[ 442.697467] env[61728]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61728) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 442.697884] env[61728]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61728) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 442.697884] env[61728]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61728) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 442.698218] env[61728]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 442.798284] env[61728]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61728) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 442.810533] env[61728]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.012s {{(pid=61728) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 443.412785] env[61728]: INFO nova.virt.driver [None req-a4da9feb-40e4-4756-99f0-9b1bb9cb7f92 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 443.485134] env[61728]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 443.485289] env[61728]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 443.485408] env[61728]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61728) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 446.634282] env[61728]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-ab9b3b68-94cd-4e60-9cd8-c5f495edfd2a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 446.650587] env[61728]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61728) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 446.650693] env[61728]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-36e08dac-925b-4dc6-b6e0-d5fb309c4b9c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 446.675341] env[61728]: INFO oslo_vmware.api [-] Successfully established new session; session ID is b55fd. [ 446.675491] env[61728]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.190s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 446.675953] env[61728]: INFO nova.virt.vmwareapi.driver [None req-a4da9feb-40e4-4756-99f0-9b1bb9cb7f92 None None] VMware vCenter version: 7.0.3 [ 446.679409] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f87cc1b-ccec-4a51-9140-62ab6840401e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 446.696988] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e03712-d092-4ded-980a-b6f14ac58ebb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 446.702527] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ff0f82-98db-4079-bacb-72c34a311057 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 446.708897] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eab8b92-babb-4221-bf31-11bc1aad3183 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 446.721748] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce84b73-de18-4951-806b-b66f80b54544 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 446.727378] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d4cfa2-3662-49b4-9c02-3ba672f3d615 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 446.756699] env[61728]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-972305a7-667c-4959-80a0-3a98c4b7ed73 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 446.761505] env[61728]: DEBUG nova.virt.vmwareapi.driver [None req-a4da9feb-40e4-4756-99f0-9b1bb9cb7f92 None None] Extension org.openstack.compute already exists. {{(pid=61728) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:224}} [ 446.764169] env[61728]: INFO nova.compute.provider_config [None req-a4da9feb-40e4-4756-99f0-9b1bb9cb7f92 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 447.267095] env[61728]: DEBUG nova.context [None req-a4da9feb-40e4-4756-99f0-9b1bb9cb7f92 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),4d3d93ae-f901-4d5a-af55-0af0918189db(cell1) {{(pid=61728) load_cells /opt/stack/nova/nova/context.py:464}} [ 447.270260] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 447.270492] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 447.271211] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 447.271769] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Acquiring lock "4d3d93ae-f901-4d5a-af55-0af0918189db" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 447.271967] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Lock "4d3d93ae-f901-4d5a-af55-0af0918189db" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 447.273086] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Lock "4d3d93ae-f901-4d5a-af55-0af0918189db" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 447.292919] env[61728]: INFO dbcounter [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Registered counter for database nova_cell0 [ 447.300907] env[61728]: INFO dbcounter [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Registered counter for database nova_cell1 [ 447.304041] env[61728]: DEBUG oslo_db.sqlalchemy.engines [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61728) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 447.304661] env[61728]: DEBUG oslo_db.sqlalchemy.engines [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61728) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 447.309211] env[61728]: ERROR nova.db.main.api [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 447.309211] env[61728]: result = function(*args, **kwargs) [ 447.309211] env[61728]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 447.309211] env[61728]: return func(*args, **kwargs) [ 447.309211] env[61728]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 447.309211] env[61728]: result = fn(*args, **kwargs) [ 447.309211] env[61728]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 447.309211] env[61728]: return f(*args, **kwargs) [ 447.309211] env[61728]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 447.309211] env[61728]: return db.service_get_minimum_version(context, binaries) [ 447.309211] env[61728]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 447.309211] env[61728]: _check_db_access() [ 447.309211] env[61728]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 447.309211] env[61728]: stacktrace = ''.join(traceback.format_stack()) [ 447.309211] env[61728]: [ 447.310045] env[61728]: ERROR nova.db.main.api [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 447.310045] env[61728]: result = function(*args, **kwargs) [ 447.310045] env[61728]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 447.310045] env[61728]: return func(*args, **kwargs) [ 447.310045] env[61728]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 447.310045] env[61728]: result = fn(*args, **kwargs) [ 447.310045] env[61728]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 447.310045] env[61728]: return f(*args, **kwargs) [ 447.310045] env[61728]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 447.310045] env[61728]: return db.service_get_minimum_version(context, binaries) [ 447.310045] env[61728]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 447.310045] env[61728]: _check_db_access() [ 447.310045] env[61728]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 447.310045] env[61728]: stacktrace = ''.join(traceback.format_stack()) [ 447.310045] env[61728]: [ 447.310446] env[61728]: WARNING nova.objects.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 447.310563] env[61728]: WARNING nova.objects.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Failed to get minimum service version for cell 4d3d93ae-f901-4d5a-af55-0af0918189db [ 447.310987] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Acquiring lock "singleton_lock" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 447.311164] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Acquired lock "singleton_lock" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 447.311407] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Releasing lock "singleton_lock" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 447.311749] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Full set of CONF: {{(pid=61728) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 447.311898] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ******************************************************************************** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 447.312030] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] Configuration options gathered from: {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 447.312171] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 447.312361] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 447.312492] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ================================================================================ {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 447.312704] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] allow_resize_to_same_host = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.312875] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] arq_binding_timeout = 300 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.313016] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] backdoor_port = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.313153] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] backdoor_socket = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.313321] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] block_device_allocate_retries = 60 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.313485] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] block_device_allocate_retries_interval = 3 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.313657] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cert = self.pem {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.313828] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.313998] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute_monitors = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.314185] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] config_dir = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.314359] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] config_drive_format = iso9660 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.314494] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.314663] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] config_source = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.314833] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] console_host = devstack {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.315007] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] control_exchange = nova {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.315178] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cpu_allocation_ratio = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.315340] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] daemon = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.315508] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] debug = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.315667] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] default_access_ip_network_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.315832] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] default_availability_zone = nova {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.315988] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] default_ephemeral_format = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.316173] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] default_green_pool_size = 1000 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.316413] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.316643] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] default_schedule_zone = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.316733] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] disk_allocation_ratio = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.316897] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] enable_new_services = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.317095] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] enabled_apis = ['osapi_compute'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.317267] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] enabled_ssl_apis = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.317428] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] flat_injected = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.317587] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] force_config_drive = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.317748] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] force_raw_images = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.317920] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] graceful_shutdown_timeout = 5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.318094] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] heal_instance_info_cache_interval = 60 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.318314] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] host = cpu-1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.318490] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.318661] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.318825] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.319050] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.319221] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] instance_build_timeout = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.319405] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] instance_delete_interval = 300 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.319596] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] instance_format = [instance: %(uuid)s] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.319770] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] instance_name_template = instance-%08x {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.319938] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] instance_usage_audit = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.320134] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] instance_usage_audit_period = month {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.320310] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.320479] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.320654] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] internal_service_availability_zone = internal {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.320816] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] key = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.320980] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] live_migration_retry_count = 30 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.321167] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] log_color = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.321334] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] log_config_append = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.321523] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.321696] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] log_dir = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.321859] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] log_file = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.321991] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] log_options = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.322172] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] log_rotate_interval = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.322343] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] log_rotate_interval_type = days {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.322512] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] log_rotation_type = none {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.322651] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.322781] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.322950] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.323131] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.323263] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.323430] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] long_rpc_timeout = 1800 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.323593] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] max_concurrent_builds = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.323757] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] max_concurrent_live_migrations = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.323920] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] max_concurrent_snapshots = 5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.324094] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] max_local_block_devices = 3 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.324262] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] max_logfile_count = 30 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.324424] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] max_logfile_size_mb = 200 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.324587] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] maximum_instance_delete_attempts = 5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.324759] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] metadata_listen = 0.0.0.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.324928] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] metadata_listen_port = 8775 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.325113] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] metadata_workers = 2 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.325281] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] migrate_max_retries = -1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.325457] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] mkisofs_cmd = genisoimage {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.325675] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.325811] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] my_ip = 10.180.1.21 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.325978] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] network_allocate_retries = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.326180] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.326359] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.326527] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] osapi_compute_listen_port = 8774 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.326730] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] osapi_compute_unique_server_name_scope = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.326871] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] osapi_compute_workers = 2 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.327048] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] password_length = 12 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.327217] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] periodic_enable = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.327379] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] periodic_fuzzy_delay = 60 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.327552] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] pointer_model = usbtablet {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.327716] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] preallocate_images = none {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.327879] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] publish_errors = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.328024] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] pybasedir = /opt/stack/nova {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.328180] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ram_allocation_ratio = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.328346] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] rate_limit_burst = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.328517] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] rate_limit_except_level = CRITICAL {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.328686] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] rate_limit_interval = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.328849] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] reboot_timeout = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.329016] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] reclaim_instance_interval = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.329180] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] record = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.329357] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] reimage_timeout_per_gb = 60 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.329550] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] report_interval = 120 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.329720] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] rescue_timeout = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.329884] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] reserved_host_cpus = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.330059] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] reserved_host_disk_mb = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.330225] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] reserved_host_memory_mb = 512 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.330390] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] reserved_huge_pages = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.330553] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] resize_confirm_window = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.330718] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] resize_fs_using_block_device = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.330879] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] resume_guests_state_on_host_boot = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.331058] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.331227] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] rpc_response_timeout = 60 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.331396] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] run_external_periodic_tasks = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.331606] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] running_deleted_instance_action = reap {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.331777] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.331942] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] running_deleted_instance_timeout = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.332120] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] scheduler_instance_sync_interval = 120 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.332294] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] service_down_time = 720 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.332465] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] servicegroup_driver = db {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.332627] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] shell_completion = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.332788] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] shelved_offload_time = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.332950] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] shelved_poll_interval = 3600 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.333134] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] shutdown_timeout = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.333300] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] source_is_ipv6 = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.333460] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ssl_only = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.333709] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.333881] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] sync_power_state_interval = 600 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.334056] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] sync_power_state_pool_size = 1000 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.334232] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] syslog_log_facility = LOG_USER {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.334393] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] tempdir = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.334560] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] timeout_nbd = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.334731] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] transport_url = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.334896] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] update_resources_interval = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.335069] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] use_cow_images = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.335234] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] use_eventlog = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.335395] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] use_journal = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.335554] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] use_json = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.335716] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] use_rootwrap_daemon = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.335875] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] use_stderr = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.336044] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] use_syslog = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.336207] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vcpu_pin_set = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.336375] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vif_plugging_is_fatal = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.336542] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vif_plugging_timeout = 300 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.336711] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] virt_mkfs = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.336880] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] volume_usage_poll_interval = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.337053] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] watch_log_file = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.337230] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] web = /usr/share/spice-html5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.337416] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_concurrency.disable_process_locking = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.337700] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.337888] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.338070] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.338250] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.338454] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.338591] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.338779] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.auth_strategy = keystone {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.338949] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.compute_link_prefix = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.339143] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.339325] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.dhcp_domain = novalocal {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.339530] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.enable_instance_password = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.339705] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.glance_link_prefix = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.339876] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.340064] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.340235] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.instance_list_per_project_cells = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.340404] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.list_records_by_skipping_down_cells = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.340572] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.local_metadata_per_cell = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.340754] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.max_limit = 1000 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.340926] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.metadata_cache_expiration = 15 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.341119] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.neutron_default_tenant_id = default {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.341296] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.response_validation = warn {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.341486] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.use_neutron_default_nets = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.341675] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.341846] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.342030] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.342215] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.342391] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.vendordata_dynamic_targets = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.342581] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.vendordata_jsonfile_path = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.342774] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.342971] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.backend = dogpile.cache.memcached {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.343158] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.backend_argument = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.343337] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.config_prefix = cache.oslo {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.343539] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.dead_timeout = 60.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.343729] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.debug_cache_backend = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.343900] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.enable_retry_client = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.344078] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.enable_socket_keepalive = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.344258] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.enabled = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.344427] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.enforce_fips_mode = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.344596] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.expiration_time = 600 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.344766] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.hashclient_retry_attempts = 2 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.344938] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.345119] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.memcache_dead_retry = 300 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.345287] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.memcache_password = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.345457] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.345628] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.345794] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.memcache_pool_maxsize = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.345960] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.346140] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.memcache_sasl_enabled = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.346322] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.346496] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.346660] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.memcache_username = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.346831] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.proxies = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.346998] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.redis_db = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.347174] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.redis_password = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.347350] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.347527] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.347701] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.redis_server = localhost:6379 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.347871] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.redis_socket_timeout = 1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.348047] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.redis_username = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.348217] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.retry_attempts = 2 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.348388] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.retry_delay = 0.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.348557] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.socket_keepalive_count = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.348725] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.socket_keepalive_idle = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.348890] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.socket_keepalive_interval = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.349070] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.tls_allowed_ciphers = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.349236] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.tls_cafile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.349414] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.tls_certfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.349595] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.tls_enabled = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.349760] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cache.tls_keyfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.349931] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.auth_section = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.350123] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.auth_type = password {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.350291] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.cafile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.350473] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.350664] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.certfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.350864] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.collect_timing = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.351047] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.cross_az_attach = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.351220] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.debug = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.351383] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.endpoint_template = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.351579] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.http_retries = 3 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.351755] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.insecure = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.351919] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.keyfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.352111] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.os_region_name = RegionOne {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.352284] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.split_loggers = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.352451] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cinder.timeout = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.352627] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.352794] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute.cpu_dedicated_set = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.352957] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute.cpu_shared_set = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.353143] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute.image_type_exclude_list = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.353312] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.353480] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.353647] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.353812] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.353984] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.354218] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute.resource_provider_association_refresh = 300 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.354329] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.354494] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute.shutdown_retry_interval = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.354679] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.354860] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] conductor.workers = 2 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.355053] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] console.allowed_origins = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.355223] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] console.ssl_ciphers = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.355397] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] console.ssl_minimum_version = default {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.355593] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] consoleauth.enforce_session_timeout = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.355777] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] consoleauth.token_ttl = 600 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.355954] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.cafile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.356129] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.certfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.356298] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.collect_timing = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.356460] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.connect_retries = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.356624] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.connect_retry_delay = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.356784] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.endpoint_override = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.356951] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.insecure = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.357126] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.keyfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.357292] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.max_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.357455] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.min_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.357616] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.region_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.357785] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.retriable_status_codes = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.357935] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.service_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.358119] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.service_type = accelerator {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.358287] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.split_loggers = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.358447] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.status_code_retries = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.358610] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.status_code_retry_delay = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.358772] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.timeout = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.358954] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.359134] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] cyborg.version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.359320] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.backend = sqlalchemy {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.359523] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.connection = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.359699] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.connection_debug = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.359873] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.connection_parameters = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.360053] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.connection_recycle_time = 3600 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.360223] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.connection_trace = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.360387] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.db_inc_retry_interval = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.360552] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.db_max_retries = 20 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.360717] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.db_max_retry_interval = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.360882] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.db_retry_interval = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.361056] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.max_overflow = 50 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.361224] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.max_pool_size = 5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.361388] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.max_retries = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.361593] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.361759] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.mysql_wsrep_sync_wait = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.361933] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.pool_timeout = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.362496] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.retry_interval = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.362496] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.slave_connection = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.362496] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.sqlite_synchronous = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.362663] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] database.use_db_reconnect = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.362828] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.backend = sqlalchemy {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.363013] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.connection = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.363191] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.connection_debug = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.363363] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.connection_parameters = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.363531] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.connection_recycle_time = 3600 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.363699] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.connection_trace = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.363872] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.db_inc_retry_interval = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.364047] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.db_max_retries = 20 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.364217] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.db_max_retry_interval = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.364379] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.db_retry_interval = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.364541] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.max_overflow = 50 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.364707] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.max_pool_size = 5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.364870] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.max_retries = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.365054] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.365221] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.365380] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.pool_timeout = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.365544] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.retry_interval = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.365705] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.slave_connection = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.365867] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] api_database.sqlite_synchronous = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.366056] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] devices.enabled_mdev_types = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.366242] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.366423] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.366585] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ephemeral_storage_encryption.enabled = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.366752] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.366925] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.api_servers = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.367108] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.cafile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.367278] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.certfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.367443] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.collect_timing = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.367635] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.connect_retries = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.367811] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.connect_retry_delay = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.367979] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.debug = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.368162] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.default_trusted_certificate_ids = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.368328] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.enable_certificate_validation = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.368490] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.enable_rbd_download = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.368653] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.endpoint_override = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.368820] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.insecure = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.368985] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.keyfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.369161] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.max_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.369323] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.min_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.369491] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.num_retries = 3 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.369662] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.rbd_ceph_conf = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.369828] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.rbd_connect_timeout = 5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.369998] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.rbd_pool = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.370187] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.rbd_user = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.370352] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.region_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.370512] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.retriable_status_codes = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.370675] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.service_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.370844] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.service_type = image {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.371013] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.split_loggers = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.371183] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.status_code_retries = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.371336] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.status_code_retry_delay = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.371516] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.timeout = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.371718] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.371886] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.verify_glance_signatures = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.372059] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] glance.version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.372233] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] guestfs.debug = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.372400] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] mks.enabled = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.372780] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.372980] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] image_cache.manager_interval = 2400 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.373171] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] image_cache.precache_concurrency = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.373344] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] image_cache.remove_unused_base_images = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.373520] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.373694] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.373874] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] image_cache.subdirectory_name = _base {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.374075] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.api_max_retries = 60 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.374252] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.api_retry_interval = 2 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.374418] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.auth_section = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.374617] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.auth_type = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.374805] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.cafile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.374970] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.certfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.375153] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.collect_timing = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.375321] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.conductor_group = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.375487] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.connect_retries = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.375651] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.connect_retry_delay = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.375813] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.endpoint_override = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.375976] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.insecure = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.376151] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.keyfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.376314] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.max_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.376476] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.min_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.376646] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.peer_list = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.376809] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.region_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.376972] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.retriable_status_codes = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.377154] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.serial_console_state_timeout = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.377317] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.service_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.377491] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.service_type = baremetal {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.377654] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.shard = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.377820] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.split_loggers = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.377983] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.status_code_retries = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.378160] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.status_code_retry_delay = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.378321] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.timeout = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.378506] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.378671] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ironic.version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.378856] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.379043] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] key_manager.fixed_key = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.379237] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.379438] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.barbican_api_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.379647] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.barbican_endpoint = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.379837] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.barbican_endpoint_type = public {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.380009] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.barbican_region_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.380185] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.cafile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.380347] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.certfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.380514] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.collect_timing = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.380679] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.insecure = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.380839] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.keyfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.381013] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.number_of_retries = 60 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.381185] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.retry_delay = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.381352] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.send_service_user_token = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.381539] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.split_loggers = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.381713] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.timeout = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.381879] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.verify_ssl = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.382052] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican.verify_ssl_path = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.382229] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican_service_user.auth_section = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.382400] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican_service_user.auth_type = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.382582] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican_service_user.cafile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.382750] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican_service_user.certfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.382918] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican_service_user.collect_timing = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.383102] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican_service_user.insecure = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.383279] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican_service_user.keyfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.383447] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican_service_user.split_loggers = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.383612] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] barbican_service_user.timeout = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.383782] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vault.approle_role_id = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.383944] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vault.approle_secret_id = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.384133] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vault.kv_mountpoint = secret {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.384297] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vault.kv_path = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.384464] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vault.kv_version = 2 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.384627] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vault.namespace = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.384788] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vault.root_token_id = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.384948] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vault.ssl_ca_crt_file = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.385131] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vault.timeout = 60.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.385297] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vault.use_ssl = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.385467] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.385644] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.auth_section = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.385808] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.auth_type = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.385968] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.cafile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.386141] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.certfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.386307] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.collect_timing = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.386467] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.connect_retries = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.386667] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.connect_retry_delay = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.386882] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.endpoint_override = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.387068] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.insecure = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.387236] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.keyfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.387397] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.max_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.387558] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.min_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.387720] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.region_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.387879] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.retriable_status_codes = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.388049] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.service_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.388227] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.service_type = identity {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.388391] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.split_loggers = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.388551] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.status_code_retries = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.388714] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.status_code_retry_delay = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.388874] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.timeout = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.389071] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.389239] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] keystone.version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.389468] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.connection_uri = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.389644] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.cpu_mode = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.389815] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.389986] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.cpu_models = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.390175] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.cpu_power_governor_high = performance {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.390347] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.390513] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.cpu_power_management = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.390690] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.390857] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.device_detach_attempts = 8 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.391033] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.device_detach_timeout = 20 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.391206] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.disk_cachemodes = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.391367] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.disk_prefix = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.391571] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.enabled_perf_events = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.391759] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.file_backed_memory = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.391931] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.gid_maps = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.392110] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.hw_disk_discard = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.392275] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.hw_machine_type = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.392448] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.images_rbd_ceph_conf = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.392617] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.392784] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.392960] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.images_rbd_glance_store_name = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.393154] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.images_rbd_pool = rbd {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.393331] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.images_type = default {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.393495] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.images_volume_group = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.393663] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.inject_key = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.393827] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.inject_partition = -2 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.393989] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.inject_password = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.394168] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.iscsi_iface = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.394331] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.iser_use_multipath = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.394499] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.394667] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.394833] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.live_migration_downtime = 500 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.394999] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.395178] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.395339] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.live_migration_inbound_addr = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.395503] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.395668] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.395831] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.live_migration_scheme = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.396010] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.live_migration_timeout_action = abort {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.396189] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.live_migration_tunnelled = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.396355] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.live_migration_uri = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.396527] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.live_migration_with_native_tls = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.396694] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.max_queues = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.396864] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.397110] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.397282] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.nfs_mount_options = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.398461] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.398659] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.398837] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.399013] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.399197] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.399383] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.num_pcie_ports = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.399652] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.399850] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.pmem_namespaces = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.400031] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.quobyte_client_cfg = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.400329] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.400509] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.400687] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.400858] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.401036] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.rbd_secret_uuid = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.401207] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.rbd_user = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.401377] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.401584] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.401765] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.rescue_image_id = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.401931] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.rescue_kernel_id = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.402108] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.rescue_ramdisk_id = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.402288] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.402454] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.rx_queue_size = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.402646] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.smbfs_mount_options = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.402939] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.403133] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.snapshot_compression = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.403301] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.snapshot_image_format = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.403526] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.403737] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.sparse_logical_volumes = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.403910] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.swtpm_enabled = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.404100] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.swtpm_group = tss {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.404280] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.swtpm_user = tss {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.404458] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.sysinfo_serial = unique {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.404628] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.tb_cache_size = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.404794] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.tx_queue_size = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.404963] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.uid_maps = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.405146] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.use_virtio_for_bridges = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.405325] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.virt_type = kvm {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.405499] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.volume_clear = zero {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.405670] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.volume_clear_size = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.405842] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.volume_use_multipath = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.406014] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.vzstorage_cache_path = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.406197] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.406373] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.406542] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.406718] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.407008] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.407204] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.vzstorage_mount_user = stack {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.407378] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.407559] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.auth_section = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.407743] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.auth_type = password {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.407911] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.cafile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.408089] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.certfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.408263] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.collect_timing = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.408425] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.connect_retries = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.408592] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.connect_retry_delay = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.408767] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.default_floating_pool = public {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.408932] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.endpoint_override = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.409113] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.extension_sync_interval = 600 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.409284] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.http_retries = 3 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.409475] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.insecure = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.409648] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.keyfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.409813] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.max_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.409986] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.410166] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.min_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.410341] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.ovs_bridge = br-int {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.410511] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.physnets = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.410685] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.region_name = RegionOne {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.410849] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.retriable_status_codes = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.411031] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.service_metadata_proxy = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.411200] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.service_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.411374] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.service_type = network {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.411564] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.split_loggers = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.411738] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.status_code_retries = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.411905] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.status_code_retry_delay = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.412080] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.timeout = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.412272] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.412442] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] neutron.version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.412634] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] notifications.bdms_in_notifications = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.412830] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] notifications.default_level = INFO {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.413020] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] notifications.notification_format = unversioned {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.413195] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] notifications.notify_on_state_change = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.413378] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.413560] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] pci.alias = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.413740] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] pci.device_spec = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.413908] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] pci.report_in_placement = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.414101] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.auth_section = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.414284] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.auth_type = password {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.414460] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.414629] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.cafile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.414791] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.certfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.414957] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.collect_timing = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.415137] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.connect_retries = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.415305] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.connect_retry_delay = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.415471] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.default_domain_id = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.415661] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.default_domain_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.415845] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.domain_id = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.416018] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.domain_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.416188] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.endpoint_override = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.416357] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.insecure = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.416520] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.keyfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.416685] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.max_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.416847] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.min_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.417033] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.password = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.417205] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.project_domain_id = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.417376] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.project_domain_name = Default {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.417548] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.project_id = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.417726] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.project_name = service {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.417900] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.region_name = RegionOne {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.418078] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.retriable_status_codes = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.418247] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.service_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.418421] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.service_type = placement {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.418592] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.split_loggers = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.418759] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.status_code_retries = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.418926] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.status_code_retry_delay = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.419104] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.system_scope = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.419274] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.timeout = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.419463] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.trust_id = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.419639] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.user_domain_id = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.419813] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.user_domain_name = Default {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.419978] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.user_id = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.420173] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.username = nova {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.420362] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.420531] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] placement.version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.420716] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] quota.cores = 20 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.420883] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] quota.count_usage_from_placement = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.421069] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.421246] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] quota.injected_file_content_bytes = 10240 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.421419] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] quota.injected_file_path_length = 255 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.421617] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] quota.injected_files = 5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.421797] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] quota.instances = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.421969] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] quota.key_pairs = 100 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.422155] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] quota.metadata_items = 128 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.422326] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] quota.ram = 51200 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.422494] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] quota.recheck_quota = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.422691] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] quota.server_group_members = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.422867] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] quota.server_groups = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.423057] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.423232] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.423392] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] scheduler.image_metadata_prefilter = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.423555] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.423730] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] scheduler.max_attempts = 3 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.423897] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] scheduler.max_placement_results = 1000 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.424073] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.424241] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.424404] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.424579] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] scheduler.workers = 2 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.424761] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.424943] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.425140] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.425316] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.425486] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.425656] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.425825] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.426023] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.426200] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.host_subset_size = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.426373] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.426539] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.426709] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.426881] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.isolated_hosts = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.427066] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.isolated_images = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.427237] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.427402] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.427569] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.427772] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.pci_in_placement = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.427954] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.428139] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.428308] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.428471] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.428641] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.428808] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.428971] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.track_instance_changes = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.429167] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.429344] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] metrics.required = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.429514] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] metrics.weight_multiplier = 1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.429684] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.429853] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] metrics.weight_setting = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.430185] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.430366] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] serial_console.enabled = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.430549] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] serial_console.port_range = 10000:20000 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.430728] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.430901] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.431084] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] serial_console.serialproxy_port = 6083 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.431262] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] service_user.auth_section = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.431438] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] service_user.auth_type = password {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.431631] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] service_user.cafile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.431799] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] service_user.certfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.431968] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] service_user.collect_timing = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.432151] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] service_user.insecure = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.432317] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] service_user.keyfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.432493] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] service_user.send_service_user_token = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.432662] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] service_user.split_loggers = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.432825] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] service_user.timeout = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.432997] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] spice.agent_enabled = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.433177] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] spice.enabled = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.433504] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.433701] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.433874] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] spice.html5proxy_port = 6082 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.434047] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] spice.image_compression = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.434215] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] spice.jpeg_compression = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.434378] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] spice.playback_compression = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.434551] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] spice.server_listen = 127.0.0.1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.434724] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.434884] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] spice.streaming_mode = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.435055] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] spice.zlib_compression = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.435225] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] upgrade_levels.baseapi = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.435396] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] upgrade_levels.compute = auto {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.435558] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] upgrade_levels.conductor = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.435719] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] upgrade_levels.scheduler = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.435888] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.436064] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.436235] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vendordata_dynamic_auth.cafile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.436398] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vendordata_dynamic_auth.certfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.436561] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.436726] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vendordata_dynamic_auth.insecure = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.436887] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.437061] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.437225] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vendordata_dynamic_auth.timeout = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.437400] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.api_retry_count = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.437564] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.ca_file = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.437740] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.437910] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.cluster_name = testcl1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.438090] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.connection_pool_size = 10 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.438255] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.console_delay_seconds = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.438427] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.datastore_regex = ^datastore.* {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.438631] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.438807] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.host_password = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.438977] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.host_port = 443 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.439164] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.host_username = administrator@vsphere.local {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.439339] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.insecure = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.439539] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.integration_bridge = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.439716] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.maximum_objects = 100 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.439882] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.pbm_default_policy = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.440059] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.pbm_enabled = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.440226] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.pbm_wsdl_location = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.440400] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.440562] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.serial_port_proxy_uri = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.440726] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.serial_port_service_uri = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.440898] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.task_poll_interval = 0.5 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.441076] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.use_linked_clone = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.441250] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.vnc_keymap = en-us {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.441419] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.vnc_port = 5900 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.441610] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vmware.vnc_port_total = 10000 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.441806] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vnc.auth_schemes = ['none'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.441985] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vnc.enabled = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.442292] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.442484] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.442660] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vnc.novncproxy_port = 6080 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.442838] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vnc.server_listen = 127.0.0.1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.443021] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.443190] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vnc.vencrypt_ca_certs = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.443352] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vnc.vencrypt_client_cert = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.443512] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vnc.vencrypt_client_key = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.443689] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.443855] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.disable_deep_image_inspection = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.444031] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.444202] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.444365] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.444529] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.disable_rootwrap = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.444695] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.enable_numa_live_migration = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.444859] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.445033] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.445203] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.445366] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.libvirt_disable_apic = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.445527] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.445696] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.445858] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.446030] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.446198] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.446362] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.446525] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.446689] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.446851] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.447028] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.447220] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.447392] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] wsgi.client_socket_timeout = 900 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.447559] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] wsgi.default_pool_size = 1000 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.447729] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] wsgi.keep_alive = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.447896] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] wsgi.max_header_line = 16384 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.448073] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.448241] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] wsgi.ssl_ca_file = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.448404] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] wsgi.ssl_cert_file = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.448563] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] wsgi.ssl_key_file = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.448731] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] wsgi.tcp_keepidle = 600 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.448909] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.449092] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] zvm.ca_file = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.449259] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] zvm.cloud_connector_url = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.449578] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.449760] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] zvm.reachable_timeout = 300 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.449946] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_policy.enforce_new_defaults = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.450135] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_policy.enforce_scope = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.450314] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_policy.policy_default_rule = default {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.450523] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.450678] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_policy.policy_file = policy.yaml {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.450849] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.451020] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.451188] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.451349] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.451559] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.451747] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.451925] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.452120] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] profiler.connection_string = messaging:// {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.452294] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] profiler.enabled = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.452480] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] profiler.es_doc_type = notification {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.452664] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] profiler.es_scroll_size = 10000 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.452839] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] profiler.es_scroll_time = 2m {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.453052] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] profiler.filter_error_trace = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.453239] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] profiler.hmac_keys = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.453416] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] profiler.sentinel_service_name = mymaster {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.453592] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] profiler.socket_timeout = 0.1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.453760] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] profiler.trace_requests = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.453926] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] profiler.trace_sqlalchemy = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.454126] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] profiler_jaeger.process_tags = {} {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.454294] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] profiler_jaeger.service_name_prefix = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.454515] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] profiler_otlp.service_name_prefix = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.454716] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] remote_debug.host = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.454885] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] remote_debug.port = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.455082] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.455254] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.455423] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.455588] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.455756] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.455919] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.456096] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.456265] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.456431] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.456603] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.456769] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.456942] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.457125] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.457302] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.457477] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.457649] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.457825] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.457996] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.458182] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.458348] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.458520] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.458688] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.458851] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.459036] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.459207] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.459392] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.459585] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.459755] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.459925] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.460108] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.ssl = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.460289] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.460464] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.460631] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.460804] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.460977] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.461159] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.461352] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.461548] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_notifications.retry = -1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.461748] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.461931] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.462122] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.auth_section = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.462293] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.auth_type = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.462467] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.cafile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.462645] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.certfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.462816] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.collect_timing = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.462978] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.connect_retries = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.463160] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.connect_retry_delay = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.463323] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.endpoint_id = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.463510] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.endpoint_override = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.463690] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.insecure = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.463854] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.keyfile = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.464028] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.max_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.464196] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.min_version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.464357] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.region_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.464518] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.retriable_status_codes = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.464680] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.service_name = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.464841] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.service_type = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.465012] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.split_loggers = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.465181] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.status_code_retries = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.465344] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.status_code_retry_delay = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.465503] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.timeout = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.465664] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.valid_interfaces = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.465824] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_limit.version = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.466014] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_reports.file_event_handler = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.466246] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.466433] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] oslo_reports.log_dir = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.466614] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.466781] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.466944] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.467132] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.467305] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.467471] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.467647] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.467811] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vif_plug_ovs_privileged.group = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.467975] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.468160] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.468330] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.468490] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] vif_plug_ovs_privileged.user = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.468664] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.468846] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.469033] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.469213] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.469411] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.469601] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.469775] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.469942] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.470141] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.470320] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_ovs.isolate_vif = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.470493] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.470657] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.470830] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.471023] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.471188] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_vif_ovs.per_port_bridge = False {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.471357] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_brick.lock_path = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.471550] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.471730] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.471906] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] privsep_osbrick.capabilities = [21] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.472082] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] privsep_osbrick.group = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.472247] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] privsep_osbrick.helper_command = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.472418] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.472607] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.472777] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] privsep_osbrick.user = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.472955] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.473134] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] nova_sys_admin.group = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.473296] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] nova_sys_admin.helper_command = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.473465] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.473633] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.473794] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] nova_sys_admin.user = None {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.473925] env[61728]: DEBUG oslo_service.service [None req-3d0c561a-fb08-4047-93f0-dfedf4538f03 None None] ******************************************************************************** {{(pid=61728) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 447.474344] env[61728]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 447.977569] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Getting list of instances from cluster (obj){ [ 447.977569] env[61728]: value = "domain-c8" [ 447.977569] env[61728]: _type = "ClusterComputeResource" [ 447.977569] env[61728]: } {{(pid=61728) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 447.978679] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726f4ec6-7073-460c-ad00-f4ab94ebfdd2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 447.987742] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Got total of 0 instances {{(pid=61728) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 447.988341] env[61728]: WARNING nova.virt.vmwareapi.driver [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 447.988803] env[61728]: INFO nova.virt.node [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Generated node identity e7ceb92f-072b-409e-b888-6fe0676b32f1 [ 447.989049] env[61728]: INFO nova.virt.node [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Wrote node identity e7ceb92f-072b-409e-b888-6fe0676b32f1 to /opt/stack/data/n-cpu-1/compute_id [ 448.491868] env[61728]: WARNING nova.compute.manager [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Compute nodes ['e7ceb92f-072b-409e-b888-6fe0676b32f1'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 449.497922] env[61728]: INFO nova.compute.manager [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 450.504116] env[61728]: WARNING nova.compute.manager [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 450.504449] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 450.504688] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 450.504847] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 450.505024] env[61728]: DEBUG nova.compute.resource_tracker [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61728) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 450.505930] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2fbfd1c-244b-45a3-8924-cc75b68c1702 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 450.514317] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c396cab4-b166-46d7-bc6b-c30df74a40a1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 450.527582] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6b7ade-f96d-4b4e-a575-900c359dc849 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 450.534315] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c2e9179-6c57-4e76-8792-7045459db15c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 450.562992] env[61728]: DEBUG nova.compute.resource_tracker [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181812MB free_disk=116GB free_vcpus=48 pci_devices=None {{(pid=61728) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 450.563194] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 450.563354] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 451.065994] env[61728]: WARNING nova.compute.resource_tracker [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] No compute node record for cpu-1:e7ceb92f-072b-409e-b888-6fe0676b32f1: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host e7ceb92f-072b-409e-b888-6fe0676b32f1 could not be found. [ 451.569756] env[61728]: INFO nova.compute.resource_tracker [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: e7ceb92f-072b-409e-b888-6fe0676b32f1 [ 453.077323] env[61728]: DEBUG nova.compute.resource_tracker [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 453.077724] env[61728]: DEBUG nova.compute.resource_tracker [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 453.229712] env[61728]: INFO nova.scheduler.client.report [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] [req-e184ff2e-b259-45f1-94aa-17c8e4ae0389] Created resource provider record via placement API for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 453.245952] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7412298d-1e39-445a-9aa2-e948d837ad5c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 453.253713] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed49cbe-d18b-42c2-8c02-f8713bb034bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 453.283534] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f223ae04-44ad-46b6-acaf-9dc892e3defa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 453.290471] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc46da90-bd92-4d1d-965b-3661deeca513 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 453.303088] env[61728]: DEBUG nova.compute.provider_tree [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 116, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 453.837378] env[61728]: DEBUG nova.scheduler.client.report [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 116, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 453.837615] env[61728]: DEBUG nova.compute.provider_tree [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 0 to 1 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 453.837760] env[61728]: DEBUG nova.compute.provider_tree [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 116, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 453.889410] env[61728]: DEBUG nova.compute.provider_tree [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 1 to 2 during operation: update_traits {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 454.394624] env[61728]: DEBUG nova.compute.resource_tracker [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 454.394966] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.831s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 454.395013] env[61728]: DEBUG nova.service [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Creating RPC server for service compute {{(pid=61728) start /opt/stack/nova/nova/service.py:186}} [ 454.408323] env[61728]: DEBUG nova.service [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] Join ServiceGroup membership for this service compute {{(pid=61728) start /opt/stack/nova/nova/service.py:203}} [ 454.408498] env[61728]: DEBUG nova.servicegroup.drivers.db [None req-d1bc0257-402f-43ee-ae0b-826fa595715b None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61728) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 458.410558] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 458.914006] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Getting list of instances from cluster (obj){ [ 458.914006] env[61728]: value = "domain-c8" [ 458.914006] env[61728]: _type = "ClusterComputeResource" [ 458.914006] env[61728]: } {{(pid=61728) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 458.915263] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044c9955-85d7-44f4-882e-43e0e118dea2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 458.923770] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Got total of 0 instances {{(pid=61728) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 458.923984] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 458.924303] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Getting list of instances from cluster (obj){ [ 458.924303] env[61728]: value = "domain-c8" [ 458.924303] env[61728]: _type = "ClusterComputeResource" [ 458.924303] env[61728]: } {{(pid=61728) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 458.925164] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6454b4b7-03bd-422b-8ead-44c8088d8452 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 458.932413] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Got total of 0 instances {{(pid=61728) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 493.877724] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 493.879072] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 494.332436] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Acquiring lock "3cd416c6-9926-45ee-afe3-d3468e4e06c2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 494.332436] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Lock "3cd416c6-9926-45ee-afe3-d3468e4e06c2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 494.381260] env[61728]: DEBUG nova.compute.manager [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 494.837235] env[61728]: DEBUG nova.compute.manager [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 494.929275] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 494.929275] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 494.929799] env[61728]: INFO nova.compute.claims [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 494.991304] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquiring lock "d75ab8eb-f15c-4e0f-8376-6d91aa26ab58" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 494.991812] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Lock "d75ab8eb-f15c-4e0f-8376-6d91aa26ab58" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 495.365856] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.502980] env[61728]: DEBUG nova.compute.manager [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 496.005259] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55d7648-42d7-4a4e-8180-8c574b937021 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.022293] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab05c92-1481-43e2-8061-26afc1d74b2f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.033492] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.056834] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39dbf910-162d-4c8e-99fe-bd7ef0bd3ae2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.065690] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64eeefec-1a46-4616-9d21-bad2f81a3d82 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.081192] env[61728]: DEBUG nova.compute.provider_tree [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 496.585112] env[61728]: DEBUG nova.scheduler.client.report [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 116, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 497.090097] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.162s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 497.090697] env[61728]: DEBUG nova.compute.manager [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 497.093552] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.728s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 497.094880] env[61728]: INFO nova.compute.claims [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 497.608350] env[61728]: DEBUG nova.compute.utils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 497.609196] env[61728]: DEBUG nova.compute.manager [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 497.616321] env[61728]: DEBUG nova.network.neutron [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 498.122725] env[61728]: DEBUG nova.compute.manager [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 498.189810] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760ff758-00b4-4a93-ac53-c23c349edfa9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.199897] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-616d0a5d-cdef-4c76-8c5c-f228d734dcf5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.234606] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7dc189-51a9-4707-a9ad-20e7ab8f5446 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.242571] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1845243-b22c-47d5-9d44-10af3de00cbe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.258240] env[61728]: DEBUG nova.compute.provider_tree [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 498.271373] env[61728]: DEBUG nova.policy [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '244030ae678b4d8588086675b6d23718', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ac86255b8224fc095ca68631066afbe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 498.274511] env[61728]: DEBUG nova.scheduler.client.report [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 116, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 498.780979] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.687s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 498.781628] env[61728]: DEBUG nova.compute.manager [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 498.785466] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.753s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 498.787203] env[61728]: INFO nova.compute.claims [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 499.133727] env[61728]: DEBUG nova.compute.manager [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 499.173742] env[61728]: DEBUG nova.virt.hardware [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 499.173742] env[61728]: DEBUG nova.virt.hardware [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 499.173742] env[61728]: DEBUG nova.virt.hardware [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 499.173742] env[61728]: DEBUG nova.virt.hardware [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 499.174150] env[61728]: DEBUG nova.virt.hardware [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 499.174423] env[61728]: DEBUG nova.virt.hardware [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 499.174879] env[61728]: DEBUG nova.virt.hardware [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 499.175223] env[61728]: DEBUG nova.virt.hardware [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 499.178116] env[61728]: DEBUG nova.virt.hardware [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 499.178116] env[61728]: DEBUG nova.virt.hardware [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 499.178116] env[61728]: DEBUG nova.virt.hardware [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 499.178116] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba69ea5-73e1-4f5e-9d90-dbee2e7b99ac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.187375] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302bd67b-7191-4010-8dcb-61e503c5a40a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.205267] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e0c778-a07f-4753-ba80-6acee06cb101 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.292427] env[61728]: DEBUG nova.compute.utils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 499.297081] env[61728]: DEBUG nova.compute.manager [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 499.297483] env[61728]: DEBUG nova.network.neutron [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 499.517549] env[61728]: DEBUG nova.policy [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1bc14b91ba0f4966a545fc979fd2842f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '79fdc9db1d00422fbcb957965ecb7aec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 499.797979] env[61728]: DEBUG nova.compute.manager [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 499.830585] env[61728]: DEBUG nova.network.neutron [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Successfully created port: 1d776ec6-69e5-4f30-8201-13c69916cca0 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 499.937176] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ad00ab-c009-4ee4-8bb7-0c31538bbd45 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.948168] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b7dad5-72e1-4ed7-9a86-4ea41597ccf2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.989407] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea8f80f-84c3-4bfc-9852-d264494d5084 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.996425] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "771341ed-8b8e-470c-9686-82650f5271b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 499.996663] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "771341ed-8b8e-470c-9686-82650f5271b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 500.007165] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe8d1fc-696e-41dd-85aa-fa9fe62e2a35 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.026479] env[61728]: DEBUG nova.compute.provider_tree [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 500.500161] env[61728]: DEBUG nova.compute.manager [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 500.531570] env[61728]: DEBUG nova.scheduler.client.report [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 116, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 500.808245] env[61728]: DEBUG nova.compute.manager [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 500.859098] env[61728]: DEBUG nova.virt.hardware [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 500.859406] env[61728]: DEBUG nova.virt.hardware [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 500.862902] env[61728]: DEBUG nova.virt.hardware [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 500.863187] env[61728]: DEBUG nova.virt.hardware [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 500.863389] env[61728]: DEBUG nova.virt.hardware [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 500.863450] env[61728]: DEBUG nova.virt.hardware [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 500.863933] env[61728]: DEBUG nova.virt.hardware [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 500.863933] env[61728]: DEBUG nova.virt.hardware [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 500.864091] env[61728]: DEBUG nova.virt.hardware [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 500.864540] env[61728]: DEBUG nova.virt.hardware [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 500.864540] env[61728]: DEBUG nova.virt.hardware [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 500.865297] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8578e059-017b-4828-8140-b2596bc909e5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.876941] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0849bf-dc96-4ac3-82b4-d767537fafdb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.033243] env[61728]: DEBUG nova.network.neutron [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Successfully created port: fd11f13f-a69b-43da-bc74-f5d74871fe0e {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 501.040533] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 501.040533] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 501.040533] env[61728]: DEBUG nova.compute.manager [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 501.044610] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.007s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 501.046111] env[61728]: INFO nova.compute.claims [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 501.551825] env[61728]: DEBUG nova.compute.utils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 501.551825] env[61728]: DEBUG nova.compute.manager [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 501.556292] env[61728]: DEBUG nova.network.neutron [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 501.683099] env[61728]: DEBUG nova.policy [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fa90ed28524347fdacc6126b82d69596', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'db60141027bf4fc7bcb9ae44ab4e77ed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 502.067024] env[61728]: DEBUG nova.compute.manager [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 502.178314] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c6493ed-a592-49b2-b8f1-5120c471cdc4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.189792] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b530e9f5-a681-4284-bc0a-ecf3b51e7655 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.223238] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f476438-9229-4afa-ab0b-b81fb6e53d9d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.231421] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c31f80a0-8e84-4313-b8da-7a393a86fa09 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.249308] env[61728]: DEBUG nova.compute.provider_tree [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 502.724841] env[61728]: DEBUG nova.network.neutron [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Successfully created port: 95c872e2-49de-44f2-bc09-b8595260a29f {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 502.742066] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Acquiring lock "08ad3f4a-12a8-48d3-8829-f6744648a95e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 502.742297] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Lock "08ad3f4a-12a8-48d3-8829-f6744648a95e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 502.754243] env[61728]: DEBUG nova.network.neutron [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Successfully updated port: 1d776ec6-69e5-4f30-8201-13c69916cca0 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 502.756860] env[61728]: DEBUG nova.scheduler.client.report [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 116, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 502.864404] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 502.864784] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 502.865046] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Starting heal instance info cache {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 502.865170] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Rebuilding the list of instances to heal {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 503.085021] env[61728]: DEBUG nova.compute.manager [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 503.118974] env[61728]: DEBUG nova.virt.hardware [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 503.119512] env[61728]: DEBUG nova.virt.hardware [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 503.119512] env[61728]: DEBUG nova.virt.hardware [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 503.119628] env[61728]: DEBUG nova.virt.hardware [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 503.119778] env[61728]: DEBUG nova.virt.hardware [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 503.119919] env[61728]: DEBUG nova.virt.hardware [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 503.124236] env[61728]: DEBUG nova.virt.hardware [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 503.124457] env[61728]: DEBUG nova.virt.hardware [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 503.124457] env[61728]: DEBUG nova.virt.hardware [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 503.124772] env[61728]: DEBUG nova.virt.hardware [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 503.124926] env[61728]: DEBUG nova.virt.hardware [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 503.125793] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72af898a-83ea-4a68-bf19-70d83d0603cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.138444] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed30c159-06ec-4d61-b55c-c19d5cc17a87 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.247568] env[61728]: DEBUG nova.compute.manager [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 503.261786] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 503.262054] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 503.263324] env[61728]: DEBUG nova.network.neutron [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 503.264240] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.220s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 503.264743] env[61728]: DEBUG nova.compute.manager [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 503.371487] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Skipping network cache update for instance because it is Building. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 503.371624] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Skipping network cache update for instance because it is Building. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 503.371764] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Skipping network cache update for instance because it is Building. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 503.371891] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Skipping network cache update for instance because it is Building. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 503.372029] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Didn't find any instances for network info cache update. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 503.372235] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.372500] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.372711] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.372899] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.373114] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.373313] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.373478] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61728) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 503.373626] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.530768] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 503.531048] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 503.711663] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "a3e9af38-5115-4337-a20e-5d6c6ec655fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 503.712146] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "a3e9af38-5115-4337-a20e-5d6c6ec655fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 503.772210] env[61728]: DEBUG nova.compute.utils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 503.776698] env[61728]: DEBUG nova.compute.manager [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 503.776890] env[61728]: DEBUG nova.network.neutron [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 503.784440] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 503.784440] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 503.786207] env[61728]: INFO nova.compute.claims [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 503.876972] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 503.885071] env[61728]: DEBUG nova.policy [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '489b3e9916b8461a845e656c7ce2bd2f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cad2b09b986d4096ad91489a99b8d27e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 503.918432] env[61728]: DEBUG nova.network.neutron [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 504.035403] env[61728]: DEBUG nova.compute.manager [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 504.216547] env[61728]: DEBUG nova.compute.manager [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 504.278078] env[61728]: DEBUG nova.compute.manager [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 504.425101] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.425220] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 504.458994] env[61728]: DEBUG nova.network.neutron [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating instance_info_cache with network_info: [{"id": "1d776ec6-69e5-4f30-8201-13c69916cca0", "address": "fa:16:3e:6c:2d:56", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.159", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d776ec6-69", "ovs_interfaceid": "1d776ec6-69e5-4f30-8201-13c69916cca0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 504.506856] env[61728]: DEBUG nova.network.neutron [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Successfully updated port: fd11f13f-a69b-43da-bc74-f5d74871fe0e {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 504.561498] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.742754] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.894679] env[61728]: DEBUG nova.network.neutron [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Successfully created port: cc02ac93-777e-4bcf-8af5-69f162374735 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 504.931112] env[61728]: DEBUG nova.compute.manager [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 504.941218] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "4247e356-ef10-41b3-bb77-106103134d59" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.941218] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "4247e356-ef10-41b3-bb77-106103134d59" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 504.967881] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Releasing lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 504.968370] env[61728]: DEBUG nova.compute.manager [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Instance network_info: |[{"id": "1d776ec6-69e5-4f30-8201-13c69916cca0", "address": "fa:16:3e:6c:2d:56", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.159", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d776ec6-69", "ovs_interfaceid": "1d776ec6-69e5-4f30-8201-13c69916cca0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 504.970693] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:2d:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '457c42cd-4ddb-4374-923e-d419b7f6eaff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1d776ec6-69e5-4f30-8201-13c69916cca0', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 504.988678] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 504.992021] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-40f884b2-ae13-401f-a886-558ef1fcf012 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.006712] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Created folder: OpenStack in parent group-v4. [ 505.006712] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Creating folder: Project (2ac86255b8224fc095ca68631066afbe). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 505.006712] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-194afebc-a12a-42ce-b181-23567b67a1d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.011303] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Acquiring lock "refresh_cache-3cd416c6-9926-45ee-afe3-d3468e4e06c2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 505.011447] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Acquired lock "refresh_cache-3cd416c6-9926-45ee-afe3-d3468e4e06c2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 505.011617] env[61728]: DEBUG nova.network.neutron [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 505.017480] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b0fec7-7293-471c-9df2-6c55c65006a9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.033203] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Created folder: Project (2ac86255b8224fc095ca68631066afbe) in parent group-v121913. [ 505.033203] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Creating folder: Instances. Parent ref: group-v121914. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 505.033203] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8faa0f33-e53f-4c77-bf7b-4c23ac2b8c00 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.042483] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Created folder: Instances in parent group-v121914. [ 505.042807] env[61728]: DEBUG oslo.service.loopingcall [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 505.046215] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 505.046963] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f2b40c3a-3080-4825-bc21-05592b530b1c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.064494] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33dacaa6-7c34-4143-86a2-ddc99f9ef5d8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.100068] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfc55ee-e193-4969-92d2-c307ac1345db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.102466] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 505.102466] env[61728]: value = "task-463394" [ 505.102466] env[61728]: _type = "Task" [ 505.102466] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 505.112256] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e877995-e647-49f7-a346-0376f107b55b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.119377] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463394, 'name': CreateVM_Task} progress is 15%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 505.134180] env[61728]: DEBUG nova.compute.provider_tree [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 505.170755] env[61728]: DEBUG nova.network.neutron [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Successfully updated port: 95c872e2-49de-44f2-bc09-b8595260a29f {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 505.289770] env[61728]: DEBUG nova.compute.manager [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 505.325455] env[61728]: DEBUG nova.virt.hardware [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 505.326196] env[61728]: DEBUG nova.virt.hardware [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 505.326443] env[61728]: DEBUG nova.virt.hardware [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 505.326666] env[61728]: DEBUG nova.virt.hardware [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 505.326819] env[61728]: DEBUG nova.virt.hardware [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 505.326973] env[61728]: DEBUG nova.virt.hardware [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 505.327243] env[61728]: DEBUG nova.virt.hardware [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 505.327750] env[61728]: DEBUG nova.virt.hardware [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 505.327977] env[61728]: DEBUG nova.virt.hardware [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 505.328161] env[61728]: DEBUG nova.virt.hardware [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 505.328407] env[61728]: DEBUG nova.virt.hardware [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 505.330076] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddfe2ea1-2624-43b0-8ae9-05146f930cec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.340261] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d72fd5-946c-4953-872a-fcb7dbe14443 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.446724] env[61728]: DEBUG nova.compute.manager [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 505.454764] env[61728]: DEBUG nova.compute.manager [req-909bf745-5ba3-428f-8bc0-09e788e7dfa8 req-eafd357f-f50f-46d6-b5c4-4c2809537e26 service nova] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Received event network-vif-plugged-1d776ec6-69e5-4f30-8201-13c69916cca0 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 505.454764] env[61728]: DEBUG oslo_concurrency.lockutils [req-909bf745-5ba3-428f-8bc0-09e788e7dfa8 req-eafd357f-f50f-46d6-b5c4-4c2809537e26 service nova] Acquiring lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.454764] env[61728]: DEBUG oslo_concurrency.lockutils [req-909bf745-5ba3-428f-8bc0-09e788e7dfa8 req-eafd357f-f50f-46d6-b5c4-4c2809537e26 service nova] Lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 505.454764] env[61728]: DEBUG oslo_concurrency.lockutils [req-909bf745-5ba3-428f-8bc0-09e788e7dfa8 req-eafd357f-f50f-46d6-b5c4-4c2809537e26 service nova] Lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 505.454764] env[61728]: DEBUG nova.compute.manager [req-909bf745-5ba3-428f-8bc0-09e788e7dfa8 req-eafd357f-f50f-46d6-b5c4-4c2809537e26 service nova] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] No waiting events found dispatching network-vif-plugged-1d776ec6-69e5-4f30-8201-13c69916cca0 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 505.455133] env[61728]: WARNING nova.compute.manager [req-909bf745-5ba3-428f-8bc0-09e788e7dfa8 req-eafd357f-f50f-46d6-b5c4-4c2809537e26 service nova] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Received unexpected event network-vif-plugged-1d776ec6-69e5-4f30-8201-13c69916cca0 for instance with vm_state building and task_state spawning. [ 505.469122] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.497618] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Acquiring lock "b86bf217-d1f8-4d37-aa87-3a2d27d70420" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.497839] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Lock "b86bf217-d1f8-4d37-aa87-3a2d27d70420" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 505.585147] env[61728]: DEBUG nova.network.neutron [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 505.612926] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463394, 'name': CreateVM_Task, 'duration_secs': 0.366334} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 505.617262] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 505.633517] env[61728]: DEBUG oslo_vmware.service [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d643bc2f-fc6b-418c-ad5b-968a6098fcd1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.641356] env[61728]: DEBUG nova.scheduler.client.report [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 116, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 505.648863] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 505.649690] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 505.649886] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 505.651323] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c806e181-a3bb-428a-86f5-33cc3a0b9847 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.656134] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 505.656134] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52592ab8-6876-7cab-7e62-5e0a488d716b" [ 505.656134] env[61728]: _type = "Task" [ 505.656134] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 505.665846] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52592ab8-6876-7cab-7e62-5e0a488d716b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 505.674119] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquiring lock "refresh_cache-d75ab8eb-f15c-4e0f-8376-6d91aa26ab58" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 505.674119] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquired lock "refresh_cache-d75ab8eb-f15c-4e0f-8376-6d91aa26ab58" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 505.674221] env[61728]: DEBUG nova.network.neutron [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 505.923963] env[61728]: DEBUG nova.compute.manager [req-2c20ef6f-afb9-41a5-9509-c07a594b513e req-e85fb9d4-6b58-4d99-8521-f3c996932fcc service nova] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Received event network-vif-plugged-fd11f13f-a69b-43da-bc74-f5d74871fe0e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 505.924263] env[61728]: DEBUG oslo_concurrency.lockutils [req-2c20ef6f-afb9-41a5-9509-c07a594b513e req-e85fb9d4-6b58-4d99-8521-f3c996932fcc service nova] Acquiring lock "3cd416c6-9926-45ee-afe3-d3468e4e06c2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.924395] env[61728]: DEBUG oslo_concurrency.lockutils [req-2c20ef6f-afb9-41a5-9509-c07a594b513e req-e85fb9d4-6b58-4d99-8521-f3c996932fcc service nova] Lock "3cd416c6-9926-45ee-afe3-d3468e4e06c2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 505.924588] env[61728]: DEBUG oslo_concurrency.lockutils [req-2c20ef6f-afb9-41a5-9509-c07a594b513e req-e85fb9d4-6b58-4d99-8521-f3c996932fcc service nova] Lock "3cd416c6-9926-45ee-afe3-d3468e4e06c2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 505.926066] env[61728]: DEBUG nova.compute.manager [req-2c20ef6f-afb9-41a5-9509-c07a594b513e req-e85fb9d4-6b58-4d99-8521-f3c996932fcc service nova] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] No waiting events found dispatching network-vif-plugged-fd11f13f-a69b-43da-bc74-f5d74871fe0e {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 505.926066] env[61728]: WARNING nova.compute.manager [req-2c20ef6f-afb9-41a5-9509-c07a594b513e req-e85fb9d4-6b58-4d99-8521-f3c996932fcc service nova] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Received unexpected event network-vif-plugged-fd11f13f-a69b-43da-bc74-f5d74871fe0e for instance with vm_state building and task_state spawning. [ 505.996046] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.000935] env[61728]: DEBUG nova.compute.manager [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 506.129181] env[61728]: DEBUG nova.network.neutron [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Updating instance_info_cache with network_info: [{"id": "fd11f13f-a69b-43da-bc74-f5d74871fe0e", "address": "fa:16:3e:e0:13:2d", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.140", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd11f13f-a6", "ovs_interfaceid": "fd11f13f-a69b-43da-bc74-f5d74871fe0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 506.154115] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 506.154374] env[61728]: DEBUG nova.compute.manager [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 506.158097] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.282s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 506.158291] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 506.158436] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61728) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 506.158758] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.597s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 506.160465] env[61728]: INFO nova.compute.claims [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 506.168627] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1e8ad7-3355-44be-9ecc-9bf83ae6e4c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.193565] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 506.194048] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 506.194415] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 506.194645] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 506.195225] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 506.197261] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05416b4-cfa5-4b05-bb62-f8ede99151ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.204741] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2c9fc1be-de86-4edc-8aba-42fd207bffbf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.219618] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e106c27a-8ebb-4bd1-9e09-af6554881695 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.227642] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9032fe-a349-46c8-9ad7-93daa9ebcde4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.236412] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 506.236633] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 506.238420] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6daccba8-850a-4bf7-aa8e-183e41e5b3b3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.276018] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181803MB free_disk=116GB free_vcpus=48 pci_devices=None {{(pid=61728) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 506.276018] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.276506] env[61728]: DEBUG nova.network.neutron [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 506.283201] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce8ca6a2-f395-441c-842d-fd88c520c019 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.290399] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 506.290399] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ab4536-b2aa-5eaf-11f6-dd293479eee4" [ 506.290399] env[61728]: _type = "Task" [ 506.290399] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 506.298427] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ab4536-b2aa-5eaf-11f6-dd293479eee4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 506.525695] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.634397] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Releasing lock "refresh_cache-3cd416c6-9926-45ee-afe3-d3468e4e06c2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 506.634737] env[61728]: DEBUG nova.compute.manager [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Instance network_info: |[{"id": "fd11f13f-a69b-43da-bc74-f5d74871fe0e", "address": "fa:16:3e:e0:13:2d", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.140", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd11f13f-a6", "ovs_interfaceid": "fd11f13f-a69b-43da-bc74-f5d74871fe0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 506.635851] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:13:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '457c42cd-4ddb-4374-923e-d419b7f6eaff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fd11f13f-a69b-43da-bc74-f5d74871fe0e', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 506.647764] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Creating folder: Project (79fdc9db1d00422fbcb957965ecb7aec). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 506.648444] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-04e50519-b91e-4338-ab27-803d6cf23d39 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.659251] env[61728]: DEBUG nova.compute.utils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 506.663938] env[61728]: DEBUG nova.compute.manager [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 506.664312] env[61728]: DEBUG nova.network.neutron [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 506.666969] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Created folder: Project (79fdc9db1d00422fbcb957965ecb7aec) in parent group-v121913. [ 506.667371] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Creating folder: Instances. Parent ref: group-v121917. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 506.672992] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3c97bf99-e999-428f-92ea-d0f120e5fef9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.675541] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "adb58292-62cd-4142-ad66-db014bcbdc49" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.675807] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "adb58292-62cd-4142-ad66-db014bcbdc49" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 506.687138] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Created folder: Instances in parent group-v121917. [ 506.687483] env[61728]: DEBUG oslo.service.loopingcall [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 506.687695] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 506.687947] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1f515d50-6b50-47aa-9b0f-dd6463582e57 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.713708] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 506.713708] env[61728]: value = "task-463397" [ 506.713708] env[61728]: _type = "Task" [ 506.713708] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 506.718316] env[61728]: DEBUG nova.network.neutron [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Updating instance_info_cache with network_info: [{"id": "95c872e2-49de-44f2-bc09-b8595260a29f", "address": "fa:16:3e:22:95:d3", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.53", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95c872e2-49", "ovs_interfaceid": "95c872e2-49de-44f2-bc09-b8595260a29f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 506.725283] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463397, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 506.801831] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Preparing fetch location {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 506.802189] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Creating directory with path [datastore1] vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70/8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 506.802491] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-faa1c8b1-d8ba-44a2-b068-5bf002848e47 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.825026] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Created directory with path [datastore1] vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70/8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 506.825026] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Fetch image to [datastore1] vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70/8b767102-1435-4827-a43b-8e2e25ec780b/tmp-sparse.vmdk {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 506.825026] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Downloading image file data 8b767102-1435-4827-a43b-8e2e25ec780b to [datastore1] vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70/8b767102-1435-4827-a43b-8e2e25ec780b/tmp-sparse.vmdk on the data store datastore1 {{(pid=61728) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 506.825814] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f695553-f08c-45f6-87d2-a0f8ae6fea6a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.833927] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08cea314-7e1f-4723-b171-91ac3cff5e0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.844138] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7f361f-e0c9-45d6-b026-13c6cf27ba33 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.878441] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0866696-bb94-467b-a99c-76ea87394210 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.885592] env[61728]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-bb3b3ff3-abe4-4f36-a03c-65ae6f38193d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.960966] env[61728]: DEBUG nova.policy [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '08f1f00e76ab40f69a7bc74cd11243d4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b06a86420cd4224b088e42b2ec12019', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 506.977520] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Downloading image file data 8b767102-1435-4827-a43b-8e2e25ec780b to the data store datastore1 {{(pid=61728) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 507.053840] env[61728]: DEBUG oslo_vmware.rw_handles [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70/8b767102-1435-4827-a43b-8e2e25ec780b/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61728) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 507.173453] env[61728]: DEBUG nova.compute.manager [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 507.220787] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Releasing lock "refresh_cache-d75ab8eb-f15c-4e0f-8376-6d91aa26ab58" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 507.221164] env[61728]: DEBUG nova.compute.manager [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Instance network_info: |[{"id": "95c872e2-49de-44f2-bc09-b8595260a29f", "address": "fa:16:3e:22:95:d3", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.53", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95c872e2-49", "ovs_interfaceid": "95c872e2-49de-44f2-bc09-b8595260a29f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 507.233995] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:95:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '457c42cd-4ddb-4374-923e-d419b7f6eaff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '95c872e2-49de-44f2-bc09-b8595260a29f', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 507.248058] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Creating folder: Project (db60141027bf4fc7bcb9ae44ab4e77ed). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 507.258200] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-de02d3fe-87f9-48e9-840d-7f034deb170d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.263804] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463397, 'name': CreateVM_Task, 'duration_secs': 0.340623} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 507.265859] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 507.267863] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 507.268036] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 507.268352] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 507.268631] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Created folder: Project (db60141027bf4fc7bcb9ae44ab4e77ed) in parent group-v121913. [ 507.268793] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Creating folder: Instances. Parent ref: group-v121920. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 507.269418] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4172a8de-96b3-4b5e-92cf-e034871e26ef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.271160] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-93ef4ac4-c1f8-4988-910f-47cafae04fe0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.283028] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Waiting for the task: (returnval){ [ 507.283028] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521ec7c0-3cd7-74a1-fc94-56639057e2fb" [ 507.283028] env[61728]: _type = "Task" [ 507.283028] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 507.285090] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Created folder: Instances in parent group-v121920. [ 507.285330] env[61728]: DEBUG oslo.service.loopingcall [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 507.287256] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 507.287581] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ecd4bc1-b750-4f37-8d7f-72e40d73b7ac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.314025] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521ec7c0-3cd7-74a1-fc94-56639057e2fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 507.319649] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 507.319649] env[61728]: value = "task-463400" [ 507.319649] env[61728]: _type = "Task" [ 507.319649] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 507.329289] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463400, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 507.462259] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65542e6b-9245-438f-b10e-d79bfe78db28 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.470869] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-127fba5a-f3c3-4ba1-b526-1062619ba068 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.512392] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f88967-46e5-4758-99a1-b575447b7170 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.522462] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0be8d4-4ab6-4e8c-8bb9-3ce2cc041625 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.539869] env[61728]: DEBUG nova.compute.provider_tree [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 507.741767] env[61728]: DEBUG oslo_vmware.rw_handles [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Completed reading data from the image iterator. {{(pid=61728) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 507.741767] env[61728]: DEBUG oslo_vmware.rw_handles [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70/8b767102-1435-4827-a43b-8e2e25ec780b/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 507.793470] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 507.793813] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 507.794112] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 507.829653] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463400, 'name': CreateVM_Task, 'duration_secs': 0.368604} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 507.829830] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 507.830573] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 507.830779] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 507.831210] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 507.831537] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34ee3efc-de5b-4fef-8312-1a38a41e088f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.836600] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 507.836600] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527dc14f-18ab-723c-a3c6-20bdb542482f" [ 507.836600] env[61728]: _type = "Task" [ 507.836600] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 507.845943] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527dc14f-18ab-723c-a3c6-20bdb542482f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 507.874831] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Downloaded image file data 8b767102-1435-4827-a43b-8e2e25ec780b to vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70/8b767102-1435-4827-a43b-8e2e25ec780b/tmp-sparse.vmdk on the data store datastore1 {{(pid=61728) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 507.878322] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Caching image {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 507.878769] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Copying Virtual Disk [datastore1] vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70/8b767102-1435-4827-a43b-8e2e25ec780b/tmp-sparse.vmdk to [datastore1] vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 507.881407] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4be793da-7283-46d9-bbc8-8554a166534f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.885043] env[61728]: DEBUG nova.network.neutron [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Successfully updated port: cc02ac93-777e-4bcf-8af5-69f162374735 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 507.894054] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 507.894054] env[61728]: value = "task-463401" [ 507.894054] env[61728]: _type = "Task" [ 507.894054] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 507.905028] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463401, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 508.045022] env[61728]: DEBUG nova.scheduler.client.report [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 116, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 508.144214] env[61728]: DEBUG nova.network.neutron [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Successfully created port: e083e0fd-486c-40fb-b54d-c58808ccf0a3 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 508.187942] env[61728]: DEBUG nova.compute.manager [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 508.224449] env[61728]: DEBUG nova.virt.hardware [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 508.224449] env[61728]: DEBUG nova.virt.hardware [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 508.224449] env[61728]: DEBUG nova.virt.hardware [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 508.224653] env[61728]: DEBUG nova.virt.hardware [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 508.224653] env[61728]: DEBUG nova.virt.hardware [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 508.224653] env[61728]: DEBUG nova.virt.hardware [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 508.224860] env[61728]: DEBUG nova.virt.hardware [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 508.225076] env[61728]: DEBUG nova.virt.hardware [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 508.225158] env[61728]: DEBUG nova.virt.hardware [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 508.225315] env[61728]: DEBUG nova.virt.hardware [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 508.226577] env[61728]: DEBUG nova.virt.hardware [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 508.226577] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e08d08-1929-4ffd-a787-65573ac0f546 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.234824] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b3221f-19c8-479c-a94e-7c7d14cc494b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.348226] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 508.348942] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 508.348942] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 508.390505] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "refresh_cache-771341ed-8b8e-470c-9686-82650f5271b4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 508.391475] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "refresh_cache-771341ed-8b8e-470c-9686-82650f5271b4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 508.391475] env[61728]: DEBUG nova.network.neutron [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 508.408668] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463401, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 508.555271] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.396s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 508.555802] env[61728]: DEBUG nova.compute.manager [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 508.558372] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.816s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 508.559859] env[61728]: INFO nova.compute.claims [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 508.730729] env[61728]: DEBUG nova.compute.manager [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Received event network-changed-1d776ec6-69e5-4f30-8201-13c69916cca0 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 508.730820] env[61728]: DEBUG nova.compute.manager [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Refreshing instance network info cache due to event network-changed-1d776ec6-69e5-4f30-8201-13c69916cca0. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 508.731279] env[61728]: DEBUG oslo_concurrency.lockutils [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] Acquiring lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 508.731939] env[61728]: DEBUG oslo_concurrency.lockutils [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] Acquired lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 508.732369] env[61728]: DEBUG nova.network.neutron [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Refreshing network info cache for port 1d776ec6-69e5-4f30-8201-13c69916cca0 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 508.798488] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "e56a59a6-fc0f-45ca-810a-76b18bae173a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 508.799056] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "e56a59a6-fc0f-45ca-810a-76b18bae173a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 508.907103] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463401, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.700577} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 508.907565] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Copied Virtual Disk [datastore1] vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70/8b767102-1435-4827-a43b-8e2e25ec780b/tmp-sparse.vmdk to [datastore1] vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 508.907937] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Deleting the datastore file [datastore1] vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70/8b767102-1435-4827-a43b-8e2e25ec780b/tmp-sparse.vmdk {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 508.908576] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d03c46d-f0f4-4ab5-a37e-e325ad7ceb74 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.915891] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 508.915891] env[61728]: value = "task-463402" [ 508.915891] env[61728]: _type = "Task" [ 508.915891] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 508.926353] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463402, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 508.935116] env[61728]: DEBUG nova.network.neutron [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 509.065337] env[61728]: DEBUG nova.compute.utils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 509.072019] env[61728]: DEBUG nova.compute.manager [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 509.072019] env[61728]: DEBUG nova.network.neutron [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 509.150101] env[61728]: DEBUG nova.network.neutron [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Updating instance_info_cache with network_info: [{"id": "cc02ac93-777e-4bcf-8af5-69f162374735", "address": "fa:16:3e:fc:57:f3", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc02ac93-77", "ovs_interfaceid": "cc02ac93-777e-4bcf-8af5-69f162374735", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 509.154793] env[61728]: DEBUG nova.policy [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0777be3ba7d5466ea41ba7ee98e1f535', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f587ea8b8fe432aafd3e0ecba054bd0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 509.432248] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463402, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027973} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 509.432912] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 509.433149] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Moving file from [datastore1] vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70/8b767102-1435-4827-a43b-8e2e25ec780b to [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b. {{(pid=61728) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 509.433579] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-146b3320-036e-4244-b787-43aa9ea50e41 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.443782] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 509.443782] env[61728]: value = "task-463403" [ 509.443782] env[61728]: _type = "Task" [ 509.443782] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 509.452393] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463403, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 509.549152] env[61728]: DEBUG nova.network.neutron [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updated VIF entry in instance network info cache for port 1d776ec6-69e5-4f30-8201-13c69916cca0. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 509.549523] env[61728]: DEBUG nova.network.neutron [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating instance_info_cache with network_info: [{"id": "1d776ec6-69e5-4f30-8201-13c69916cca0", "address": "fa:16:3e:6c:2d:56", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.159", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d776ec6-69", "ovs_interfaceid": "1d776ec6-69e5-4f30-8201-13c69916cca0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 509.570712] env[61728]: DEBUG nova.network.neutron [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Successfully created port: 89110ed6-c4a4-4f3b-b706-134ce068f320 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 509.573011] env[61728]: DEBUG nova.compute.manager [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 509.655709] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "refresh_cache-771341ed-8b8e-470c-9686-82650f5271b4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 509.656738] env[61728]: DEBUG nova.compute.manager [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Instance network_info: |[{"id": "cc02ac93-777e-4bcf-8af5-69f162374735", "address": "fa:16:3e:fc:57:f3", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc02ac93-77", "ovs_interfaceid": "cc02ac93-777e-4bcf-8af5-69f162374735", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 509.656738] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "3c4f7d53-6ba6-4e2c-b324-e651640b304b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 509.656999] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "3c4f7d53-6ba6-4e2c-b324-e651640b304b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 509.662018] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:57:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f65996a3-f865-4492-9377-cd14ec8b3aae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc02ac93-777e-4bcf-8af5-69f162374735', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 509.672203] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Creating folder: Project (cad2b09b986d4096ad91489a99b8d27e). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 509.672203] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-88d2b509-a8b1-4c46-a891-f8c3ce33b0d8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.682761] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Created folder: Project (cad2b09b986d4096ad91489a99b8d27e) in parent group-v121913. [ 509.682955] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Creating folder: Instances. Parent ref: group-v121923. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 509.683207] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1bb39e63-67f8-476a-9c42-461d993f3fe7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.692721] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Created folder: Instances in parent group-v121923. [ 509.692956] env[61728]: DEBUG oslo.service.loopingcall [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 509.693200] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 509.693416] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-df1af37e-b3b9-4993-9a34-49a9e5013946 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.717037] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 509.717037] env[61728]: value = "task-463406" [ 509.717037] env[61728]: _type = "Task" [ 509.717037] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 509.732796] env[61728]: DEBUG nova.compute.manager [req-ea54e271-601b-4b8d-ae84-78cf7247bd51 req-59e2df48-d2f9-400e-a8ed-9f2582576f75 service nova] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Received event network-changed-fd11f13f-a69b-43da-bc74-f5d74871fe0e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 509.733153] env[61728]: DEBUG nova.compute.manager [req-ea54e271-601b-4b8d-ae84-78cf7247bd51 req-59e2df48-d2f9-400e-a8ed-9f2582576f75 service nova] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Refreshing instance network info cache due to event network-changed-fd11f13f-a69b-43da-bc74-f5d74871fe0e. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 509.733495] env[61728]: DEBUG oslo_concurrency.lockutils [req-ea54e271-601b-4b8d-ae84-78cf7247bd51 req-59e2df48-d2f9-400e-a8ed-9f2582576f75 service nova] Acquiring lock "refresh_cache-3cd416c6-9926-45ee-afe3-d3468e4e06c2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 509.733758] env[61728]: DEBUG oslo_concurrency.lockutils [req-ea54e271-601b-4b8d-ae84-78cf7247bd51 req-59e2df48-d2f9-400e-a8ed-9f2582576f75 service nova] Acquired lock "refresh_cache-3cd416c6-9926-45ee-afe3-d3468e4e06c2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 509.734024] env[61728]: DEBUG nova.network.neutron [req-ea54e271-601b-4b8d-ae84-78cf7247bd51 req-59e2df48-d2f9-400e-a8ed-9f2582576f75 service nova] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Refreshing network info cache for port fd11f13f-a69b-43da-bc74-f5d74871fe0e {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 509.737622] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463406, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 509.834058] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a3c205-42ef-4762-bc66-682682e89a65 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.844800] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d323af-65fe-4037-a4c5-aa66f275edc9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.879776] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ff8e5f-2256-4fb4-8c47-491d6bb4f85f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.888760] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a45176-7935-4453-abb3-b3502f0171d3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.906114] env[61728]: DEBUG nova.compute.provider_tree [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 509.959680] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463403, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.022926} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 509.959943] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] File moved {{(pid=61728) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 509.960334] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Cleaning up location [datastore1] vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70 {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 509.960334] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Deleting the datastore file [datastore1] vmware_temp/4949e5b3-4ca7-4fd2-85c1-4fdaa57eea70 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 509.960855] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf9e8ea4-36c9-493f-8a36-4ab80cb46d7d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.967677] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 509.967677] env[61728]: value = "task-463407" [ 509.967677] env[61728]: _type = "Task" [ 509.967677] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 509.982744] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463407, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 510.052363] env[61728]: DEBUG oslo_concurrency.lockutils [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] Releasing lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 510.053026] env[61728]: DEBUG nova.compute.manager [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Received event network-vif-plugged-95c872e2-49de-44f2-bc09-b8595260a29f {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 510.054621] env[61728]: DEBUG oslo_concurrency.lockutils [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] Acquiring lock "d75ab8eb-f15c-4e0f-8376-6d91aa26ab58-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.058060] env[61728]: DEBUG oslo_concurrency.lockutils [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] Lock "d75ab8eb-f15c-4e0f-8376-6d91aa26ab58-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.058060] env[61728]: DEBUG oslo_concurrency.lockutils [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] Lock "d75ab8eb-f15c-4e0f-8376-6d91aa26ab58-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 510.058060] env[61728]: DEBUG nova.compute.manager [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] No waiting events found dispatching network-vif-plugged-95c872e2-49de-44f2-bc09-b8595260a29f {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 510.058060] env[61728]: WARNING nova.compute.manager [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Received unexpected event network-vif-plugged-95c872e2-49de-44f2-bc09-b8595260a29f for instance with vm_state building and task_state spawning. [ 510.058060] env[61728]: DEBUG nova.compute.manager [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Received event network-changed-95c872e2-49de-44f2-bc09-b8595260a29f {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 510.058266] env[61728]: DEBUG nova.compute.manager [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Refreshing instance network info cache due to event network-changed-95c872e2-49de-44f2-bc09-b8595260a29f. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 510.058266] env[61728]: DEBUG oslo_concurrency.lockutils [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] Acquiring lock "refresh_cache-d75ab8eb-f15c-4e0f-8376-6d91aa26ab58" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 510.058266] env[61728]: DEBUG oslo_concurrency.lockutils [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] Acquired lock "refresh_cache-d75ab8eb-f15c-4e0f-8376-6d91aa26ab58" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 510.058266] env[61728]: DEBUG nova.network.neutron [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Refreshing network info cache for port 95c872e2-49de-44f2-bc09-b8595260a29f {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 510.159185] env[61728]: DEBUG nova.network.neutron [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Successfully updated port: e083e0fd-486c-40fb-b54d-c58808ccf0a3 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 510.226554] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463406, 'name': CreateVM_Task, 'duration_secs': 0.342573} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 510.226554] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 510.226952] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 510.227121] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 510.227430] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 510.228046] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dde402a3-dc54-4cb0-9f1a-3a1c2f73e93c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.234927] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 510.234927] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c82b7a-33d3-4a04-6aef-e5147dd0dd48" [ 510.234927] env[61728]: _type = "Task" [ 510.234927] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 510.246435] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c82b7a-33d3-4a04-6aef-e5147dd0dd48, 'name': SearchDatastore_Task, 'duration_secs': 0.008405} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 510.247329] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 510.247329] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 510.247329] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 510.321639] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "620bb472-c36e-4c56-acdf-42e02614856b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.321892] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "620bb472-c36e-4c56-acdf-42e02614856b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.408816] env[61728]: DEBUG nova.scheduler.client.report [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 116, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 510.482980] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463407, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.042654} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 510.482980] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 510.483384] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf1c05f4-48e6-4a54-84ab-074a076c34b6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.488945] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 510.488945] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52669cb1-5993-4144-351e-0a3bda785293" [ 510.488945] env[61728]: _type = "Task" [ 510.488945] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 510.498146] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52669cb1-5993-4144-351e-0a3bda785293, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 510.551949] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Acquiring lock "8c4b7227-1a01-4400-beb8-2df27e17e329" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.552044] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Lock "8c4b7227-1a01-4400-beb8-2df27e17e329" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.587928] env[61728]: DEBUG nova.compute.manager [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 510.628933] env[61728]: DEBUG nova.virt.hardware [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 510.629192] env[61728]: DEBUG nova.virt.hardware [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 510.629420] env[61728]: DEBUG nova.virt.hardware [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 510.629889] env[61728]: DEBUG nova.virt.hardware [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 510.630115] env[61728]: DEBUG nova.virt.hardware [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 510.630312] env[61728]: DEBUG nova.virt.hardware [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 510.630729] env[61728]: DEBUG nova.virt.hardware [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 510.631042] env[61728]: DEBUG nova.virt.hardware [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 510.631215] env[61728]: DEBUG nova.virt.hardware [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 510.631556] env[61728]: DEBUG nova.virt.hardware [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 510.631922] env[61728]: DEBUG nova.virt.hardware [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 510.633529] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64313622-8d39-491c-a01a-e94d72304ce5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.644153] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d7df0e-6a32-44f1-97e2-364255e4c93c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.661465] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Acquiring lock "refresh_cache-08ad3f4a-12a8-48d3-8829-f6744648a95e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 510.661598] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Acquired lock "refresh_cache-08ad3f4a-12a8-48d3-8829-f6744648a95e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 510.661851] env[61728]: DEBUG nova.network.neutron [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 510.743593] env[61728]: DEBUG nova.network.neutron [req-ea54e271-601b-4b8d-ae84-78cf7247bd51 req-59e2df48-d2f9-400e-a8ed-9f2582576f75 service nova] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Updated VIF entry in instance network info cache for port fd11f13f-a69b-43da-bc74-f5d74871fe0e. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 510.743968] env[61728]: DEBUG nova.network.neutron [req-ea54e271-601b-4b8d-ae84-78cf7247bd51 req-59e2df48-d2f9-400e-a8ed-9f2582576f75 service nova] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Updating instance_info_cache with network_info: [{"id": "fd11f13f-a69b-43da-bc74-f5d74871fe0e", "address": "fa:16:3e:e0:13:2d", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.140", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd11f13f-a6", "ovs_interfaceid": "fd11f13f-a69b-43da-bc74-f5d74871fe0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 510.913686] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.355s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 510.914255] env[61728]: DEBUG nova.compute.manager [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 510.918827] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.450s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.920333] env[61728]: INFO nova.compute.claims [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 510.999532] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52669cb1-5993-4144-351e-0a3bda785293, 'name': SearchDatastore_Task, 'duration_secs': 0.008415} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 510.999634] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 510.999837] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c/9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 511.000106] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 511.000288] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 511.000498] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d4ece2e8-98c9-43ad-8413-2d22675152a7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.002470] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1c56eb9f-935d-4a6f-b33f-c5237d579bd3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.010221] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 511.010221] env[61728]: value = "task-463408" [ 511.010221] env[61728]: _type = "Task" [ 511.010221] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 511.011379] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 511.015032] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 511.017128] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e73496e-767f-4444-8033-9be497a1034d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.024997] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463408, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 511.026759] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Waiting for the task: (returnval){ [ 511.026759] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52be3753-436d-8133-4fc5-3a9896793f72" [ 511.026759] env[61728]: _type = "Task" [ 511.026759] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 511.037832] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52be3753-436d-8133-4fc5-3a9896793f72, 'name': SearchDatastore_Task, 'duration_secs': 0.007913} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 511.038753] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b92afcc8-032e-492b-9c3d-796bc5cb89a4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.044391] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Waiting for the task: (returnval){ [ 511.044391] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a07ba1-324d-ddb1-d927-4a48c57c112c" [ 511.044391] env[61728]: _type = "Task" [ 511.044391] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 511.052425] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a07ba1-324d-ddb1-d927-4a48c57c112c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 511.106205] env[61728]: DEBUG nova.network.neutron [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Updated VIF entry in instance network info cache for port 95c872e2-49de-44f2-bc09-b8595260a29f. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 511.106591] env[61728]: DEBUG nova.network.neutron [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Updating instance_info_cache with network_info: [{"id": "95c872e2-49de-44f2-bc09-b8595260a29f", "address": "fa:16:3e:22:95:d3", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.53", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95c872e2-49", "ovs_interfaceid": "95c872e2-49de-44f2-bc09-b8595260a29f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 511.219242] env[61728]: DEBUG nova.network.neutron [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 511.248159] env[61728]: DEBUG oslo_concurrency.lockutils [req-ea54e271-601b-4b8d-ae84-78cf7247bd51 req-59e2df48-d2f9-400e-a8ed-9f2582576f75 service nova] Releasing lock "refresh_cache-3cd416c6-9926-45ee-afe3-d3468e4e06c2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 511.420361] env[61728]: DEBUG nova.compute.utils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 511.422645] env[61728]: DEBUG nova.compute.manager [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 511.422961] env[61728]: DEBUG nova.network.neutron [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 511.473620] env[61728]: DEBUG nova.network.neutron [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Updating instance_info_cache with network_info: [{"id": "e083e0fd-486c-40fb-b54d-c58808ccf0a3", "address": "fa:16:3e:05:49:1a", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape083e0fd-48", "ovs_interfaceid": "e083e0fd-486c-40fb-b54d-c58808ccf0a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 511.520942] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463408, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507895} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 511.521265] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c/9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 511.521478] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 511.521797] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75ed8c36-6f34-455f-a901-7dc396258052 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.528872] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 511.528872] env[61728]: value = "task-463409" [ 511.528872] env[61728]: _type = "Task" [ 511.528872] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 511.540452] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463409, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 511.541569] env[61728]: DEBUG nova.network.neutron [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Successfully updated port: 89110ed6-c4a4-4f3b-b706-134ce068f320 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 511.560833] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a07ba1-324d-ddb1-d927-4a48c57c112c, 'name': SearchDatastore_Task, 'duration_secs': 0.008352} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 511.560833] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 511.561034] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3cd416c6-9926-45ee-afe3-d3468e4e06c2/3cd416c6-9926-45ee-afe3-d3468e4e06c2.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 511.563742] env[61728]: DEBUG nova.policy [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '489b3e9916b8461a845e656c7ce2bd2f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cad2b09b986d4096ad91489a99b8d27e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 511.566081] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 511.566376] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 511.566511] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1445f920-f3e7-4980-bc56-af5c7ab8eac2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.568659] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7758c21e-2b63-48fd-b52a-c77e1d1cb37c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.575660] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Waiting for the task: (returnval){ [ 511.575660] env[61728]: value = "task-463410" [ 511.575660] env[61728]: _type = "Task" [ 511.575660] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 511.585333] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 511.585333] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 511.585333] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6295190c-d93a-46cf-9326-9eddb2cdb179 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.590802] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463410, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 511.594186] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 511.594186] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529a3010-d34b-08aa-4bd6-3845603370a8" [ 511.594186] env[61728]: _type = "Task" [ 511.594186] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 511.602078] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529a3010-d34b-08aa-4bd6-3845603370a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 511.609701] env[61728]: DEBUG oslo_concurrency.lockutils [req-6b40e33c-cb99-4547-8275-d3fbf1405fdf req-5f8bc2fe-fe39-4c49-a4dc-721260caed4d service nova] Releasing lock "refresh_cache-d75ab8eb-f15c-4e0f-8376-6d91aa26ab58" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 511.926856] env[61728]: DEBUG nova.compute.manager [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 511.941207] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquiring lock "a0831461-ece1-43ee-92f6-34d7d4e673e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 511.941548] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Lock "a0831461-ece1-43ee-92f6-34d7d4e673e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.977102] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Releasing lock "refresh_cache-08ad3f4a-12a8-48d3-8829-f6744648a95e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 511.977592] env[61728]: DEBUG nova.compute.manager [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Instance network_info: |[{"id": "e083e0fd-486c-40fb-b54d-c58808ccf0a3", "address": "fa:16:3e:05:49:1a", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape083e0fd-48", "ovs_interfaceid": "e083e0fd-486c-40fb-b54d-c58808ccf0a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 511.978454] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:49:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '457c42cd-4ddb-4374-923e-d419b7f6eaff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e083e0fd-486c-40fb-b54d-c58808ccf0a3', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 511.991098] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Creating folder: Project (0b06a86420cd4224b088e42b2ec12019). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 511.995891] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-929a588f-6b31-42f0-b694-acc52c78ca92 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.013438] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Created folder: Project (0b06a86420cd4224b088e42b2ec12019) in parent group-v121913. [ 512.013438] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Creating folder: Instances. Parent ref: group-v121926. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 512.013438] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fe5e231e-9738-46da-b64c-a4c5f3c493a5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.031332] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Created folder: Instances in parent group-v121926. [ 512.031829] env[61728]: DEBUG oslo.service.loopingcall [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 512.036539] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 512.041022] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-328d2cc0-8757-4047-b8e3-4baa657fc62c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.061460] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "refresh_cache-26220c0c-f535-4315-b3bb-2e5ac4d2286e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 512.061809] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquired lock "refresh_cache-26220c0c-f535-4315-b3bb-2e5ac4d2286e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 512.062108] env[61728]: DEBUG nova.network.neutron [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 512.064273] env[61728]: DEBUG nova.compute.manager [req-38335e48-cd09-4e5a-9d9a-a27f6f0d1a35 req-69d6596d-2b22-44ab-9c26-3542168dbb7b service nova] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Received event network-vif-plugged-cc02ac93-777e-4bcf-8af5-69f162374735 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 512.064652] env[61728]: DEBUG oslo_concurrency.lockutils [req-38335e48-cd09-4e5a-9d9a-a27f6f0d1a35 req-69d6596d-2b22-44ab-9c26-3542168dbb7b service nova] Acquiring lock "771341ed-8b8e-470c-9686-82650f5271b4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.064966] env[61728]: DEBUG oslo_concurrency.lockutils [req-38335e48-cd09-4e5a-9d9a-a27f6f0d1a35 req-69d6596d-2b22-44ab-9c26-3542168dbb7b service nova] Lock "771341ed-8b8e-470c-9686-82650f5271b4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.065253] env[61728]: DEBUG oslo_concurrency.lockutils [req-38335e48-cd09-4e5a-9d9a-a27f6f0d1a35 req-69d6596d-2b22-44ab-9c26-3542168dbb7b service nova] Lock "771341ed-8b8e-470c-9686-82650f5271b4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 512.065566] env[61728]: DEBUG nova.compute.manager [req-38335e48-cd09-4e5a-9d9a-a27f6f0d1a35 req-69d6596d-2b22-44ab-9c26-3542168dbb7b service nova] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] No waiting events found dispatching network-vif-plugged-cc02ac93-777e-4bcf-8af5-69f162374735 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 512.065831] env[61728]: WARNING nova.compute.manager [req-38335e48-cd09-4e5a-9d9a-a27f6f0d1a35 req-69d6596d-2b22-44ab-9c26-3542168dbb7b service nova] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Received unexpected event network-vif-plugged-cc02ac93-777e-4bcf-8af5-69f162374735 for instance with vm_state building and task_state spawning. [ 512.066108] env[61728]: DEBUG nova.compute.manager [req-38335e48-cd09-4e5a-9d9a-a27f6f0d1a35 req-69d6596d-2b22-44ab-9c26-3542168dbb7b service nova] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Received event network-changed-cc02ac93-777e-4bcf-8af5-69f162374735 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 512.066397] env[61728]: DEBUG nova.compute.manager [req-38335e48-cd09-4e5a-9d9a-a27f6f0d1a35 req-69d6596d-2b22-44ab-9c26-3542168dbb7b service nova] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Refreshing instance network info cache due to event network-changed-cc02ac93-777e-4bcf-8af5-69f162374735. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 512.067695] env[61728]: DEBUG oslo_concurrency.lockutils [req-38335e48-cd09-4e5a-9d9a-a27f6f0d1a35 req-69d6596d-2b22-44ab-9c26-3542168dbb7b service nova] Acquiring lock "refresh_cache-771341ed-8b8e-470c-9686-82650f5271b4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 512.067695] env[61728]: DEBUG oslo_concurrency.lockutils [req-38335e48-cd09-4e5a-9d9a-a27f6f0d1a35 req-69d6596d-2b22-44ab-9c26-3542168dbb7b service nova] Acquired lock "refresh_cache-771341ed-8b8e-470c-9686-82650f5271b4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 512.067695] env[61728]: DEBUG nova.network.neutron [req-38335e48-cd09-4e5a-9d9a-a27f6f0d1a35 req-69d6596d-2b22-44ab-9c26-3542168dbb7b service nova] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Refreshing network info cache for port cc02ac93-777e-4bcf-8af5-69f162374735 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 512.073268] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463409, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063303} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 512.073268] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 512.073706] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf2d25f-e97a-4be9-b955-f9c3b08cec03 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.082730] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 512.082730] env[61728]: value = "task-463413" [ 512.082730] env[61728]: _type = "Task" [ 512.082730] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.112557] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c/9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 512.123667] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9edcef14-536a-473f-854c-6f90120b0ba3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.148541] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463410, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486555} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 512.148910] env[61728]: DEBUG nova.network.neutron [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Successfully created port: 0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 512.152226] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3cd416c6-9926-45ee-afe3-d3468e4e06c2/3cd416c6-9926-45ee-afe3-d3468e4e06c2.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 512.152226] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 512.160653] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a2ed17cb-e4c9-495a-b0e1-15a27ef5c69d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.168795] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463413, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.169695] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529a3010-d34b-08aa-4bd6-3845603370a8, 'name': SearchDatastore_Task, 'duration_secs': 0.008012} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 512.176536] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 512.176536] env[61728]: value = "task-463414" [ 512.176536] env[61728]: _type = "Task" [ 512.176536] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.177105] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99bc7cfb-de8e-4351-b401-c5b39e8ac012 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.192209] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 512.192209] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fdd687-655c-dc6b-b08a-f6253432f3be" [ 512.192209] env[61728]: _type = "Task" [ 512.192209] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.199406] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Waiting for the task: (returnval){ [ 512.199406] env[61728]: value = "task-463415" [ 512.199406] env[61728]: _type = "Task" [ 512.199406] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.199776] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463414, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.226201] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463415, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.226473] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fdd687-655c-dc6b-b08a-f6253432f3be, 'name': SearchDatastore_Task, 'duration_secs': 0.011241} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 512.229237] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 512.229553] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] d75ab8eb-f15c-4e0f-8376-6d91aa26ab58/d75ab8eb-f15c-4e0f-8376-6d91aa26ab58.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 512.230287] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 512.230467] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 512.230688] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-16116ae9-d2c8-4daa-81f2-e99169c6e874 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.233137] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23432990-3cc9-4fcd-b5a4-dece2e78ef71 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.240059] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 512.240059] env[61728]: value = "task-463416" [ 512.240059] env[61728]: _type = "Task" [ 512.240059] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.244621] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 512.244805] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 512.248146] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-559926de-0139-4b6b-8cae-514c40676818 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.254413] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463416, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.257476] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 512.257476] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522efac0-6606-b7f5-9dff-bd89708a849a" [ 512.257476] env[61728]: _type = "Task" [ 512.257476] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.265226] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522efac0-6606-b7f5-9dff-bd89708a849a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.364279] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1995eedb-5478-4242-8d10-c2cef21901ed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.372044] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d8ce1d-796d-4747-8856-52c40343fa36 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.407187] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c794ffe4-085b-4098-93ac-1446e357cbcf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.416114] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4542c9ca-c0ff-4d85-aa52-1c74a5f009ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.431167] env[61728]: DEBUG nova.compute.provider_tree [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 512.598174] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463413, 'name': CreateVM_Task, 'duration_secs': 0.346533} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 512.598376] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 512.599494] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 512.599494] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 512.599876] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 512.600042] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-340a6102-f17a-404d-80b6-78525983c307 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.605811] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Waiting for the task: (returnval){ [ 512.605811] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52db5f0d-0641-75d9-ff41-9a69facf46b8" [ 512.605811] env[61728]: _type = "Task" [ 512.605811] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.615711] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52db5f0d-0641-75d9-ff41-9a69facf46b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.634338] env[61728]: DEBUG nova.network.neutron [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 512.693158] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463414, 'name': ReconfigVM_Task, 'duration_secs': 0.302916} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 512.693158] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Reconfigured VM instance instance-00000001 to attach disk [datastore1] 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c/9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 512.693158] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e070fd4-18e7-422c-839e-f4072963d5ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.704914] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 512.704914] env[61728]: value = "task-463417" [ 512.704914] env[61728]: _type = "Task" [ 512.704914] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.731243] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463417, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.731775] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463415, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066043} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 512.731864] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 512.732826] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce2ec79-d895-492f-878d-4957b863513e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.757035] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] 3cd416c6-9926-45ee-afe3-d3468e4e06c2/3cd416c6-9926-45ee-afe3-d3468e4e06c2.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 512.765773] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df3a5ec8-a820-42bc-bd9a-fbbcf7069836 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.782747] env[61728]: DEBUG nova.compute.manager [req-e326449b-732f-4373-bd5c-0da2d3e9ae33 req-e509f8a6-2446-41c0-a0f7-349070739fce service nova] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Received event network-vif-plugged-e083e0fd-486c-40fb-b54d-c58808ccf0a3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 512.782936] env[61728]: DEBUG oslo_concurrency.lockutils [req-e326449b-732f-4373-bd5c-0da2d3e9ae33 req-e509f8a6-2446-41c0-a0f7-349070739fce service nova] Acquiring lock "08ad3f4a-12a8-48d3-8829-f6744648a95e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.783182] env[61728]: DEBUG oslo_concurrency.lockutils [req-e326449b-732f-4373-bd5c-0da2d3e9ae33 req-e509f8a6-2446-41c0-a0f7-349070739fce service nova] Lock "08ad3f4a-12a8-48d3-8829-f6744648a95e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.783415] env[61728]: DEBUG oslo_concurrency.lockutils [req-e326449b-732f-4373-bd5c-0da2d3e9ae33 req-e509f8a6-2446-41c0-a0f7-349070739fce service nova] Lock "08ad3f4a-12a8-48d3-8829-f6744648a95e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 512.783711] env[61728]: DEBUG nova.compute.manager [req-e326449b-732f-4373-bd5c-0da2d3e9ae33 req-e509f8a6-2446-41c0-a0f7-349070739fce service nova] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] No waiting events found dispatching network-vif-plugged-e083e0fd-486c-40fb-b54d-c58808ccf0a3 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 512.783870] env[61728]: WARNING nova.compute.manager [req-e326449b-732f-4373-bd5c-0da2d3e9ae33 req-e509f8a6-2446-41c0-a0f7-349070739fce service nova] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Received unexpected event network-vif-plugged-e083e0fd-486c-40fb-b54d-c58808ccf0a3 for instance with vm_state building and task_state spawning. [ 512.784116] env[61728]: DEBUG nova.compute.manager [req-e326449b-732f-4373-bd5c-0da2d3e9ae33 req-e509f8a6-2446-41c0-a0f7-349070739fce service nova] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Received event network-changed-e083e0fd-486c-40fb-b54d-c58808ccf0a3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 512.784323] env[61728]: DEBUG nova.compute.manager [req-e326449b-732f-4373-bd5c-0da2d3e9ae33 req-e509f8a6-2446-41c0-a0f7-349070739fce service nova] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Refreshing instance network info cache due to event network-changed-e083e0fd-486c-40fb-b54d-c58808ccf0a3. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 512.784569] env[61728]: DEBUG oslo_concurrency.lockutils [req-e326449b-732f-4373-bd5c-0da2d3e9ae33 req-e509f8a6-2446-41c0-a0f7-349070739fce service nova] Acquiring lock "refresh_cache-08ad3f4a-12a8-48d3-8829-f6744648a95e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 512.784759] env[61728]: DEBUG oslo_concurrency.lockutils [req-e326449b-732f-4373-bd5c-0da2d3e9ae33 req-e509f8a6-2446-41c0-a0f7-349070739fce service nova] Acquired lock "refresh_cache-08ad3f4a-12a8-48d3-8829-f6744648a95e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 512.784961] env[61728]: DEBUG nova.network.neutron [req-e326449b-732f-4373-bd5c-0da2d3e9ae33 req-e509f8a6-2446-41c0-a0f7-349070739fce service nova] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Refreshing network info cache for port e083e0fd-486c-40fb-b54d-c58808ccf0a3 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 512.799986] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522efac0-6606-b7f5-9dff-bd89708a849a, 'name': SearchDatastore_Task, 'duration_secs': 0.008965} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 512.805086] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463416, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.805086] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Waiting for the task: (returnval){ [ 512.805086] env[61728]: value = "task-463418" [ 512.805086] env[61728]: _type = "Task" [ 512.805086] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.805086] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e25bc0a7-31ec-426d-880e-2dbd70eb349b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.816201] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 512.816201] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c38b8b-d2ec-24bf-ac5c-50d255e99fab" [ 512.816201] env[61728]: _type = "Task" [ 512.816201] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.818046] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463418, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.830066] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c38b8b-d2ec-24bf-ac5c-50d255e99fab, 'name': SearchDatastore_Task, 'duration_secs': 0.010567} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 512.830394] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 512.830693] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4/771341ed-8b8e-470c-9686-82650f5271b4.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 512.831050] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-764585b6-b114-4dac-a8eb-fe07692ba15e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.838560] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 512.838560] env[61728]: value = "task-463419" [ 512.838560] env[61728]: _type = "Task" [ 512.838560] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.847045] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463419, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.940411] env[61728]: DEBUG nova.compute.manager [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 512.954037] env[61728]: ERROR nova.scheduler.client.report [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [req-d9e2ed83-0ef4-4f48-9d4a-d52f34c749e8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d9e2ed83-0ef4-4f48-9d4a-d52f34c749e8"}]} [ 512.966304] env[61728]: DEBUG nova.virt.hardware [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 512.966637] env[61728]: DEBUG nova.virt.hardware [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 512.966717] env[61728]: DEBUG nova.virt.hardware [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 512.966873] env[61728]: DEBUG nova.virt.hardware [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 512.968185] env[61728]: DEBUG nova.virt.hardware [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 512.968444] env[61728]: DEBUG nova.virt.hardware [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 512.968673] env[61728]: DEBUG nova.virt.hardware [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 512.968840] env[61728]: DEBUG nova.virt.hardware [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 512.969036] env[61728]: DEBUG nova.virt.hardware [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 512.969193] env[61728]: DEBUG nova.virt.hardware [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 512.969415] env[61728]: DEBUG nova.virt.hardware [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 512.970300] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024c0d4f-f0c6-4ebf-93b1-fd085cfe95cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.978650] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb010b74-b4cf-4a50-9916-06293b155e99 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.985146] env[61728]: DEBUG nova.scheduler.client.report [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 513.004125] env[61728]: DEBUG nova.scheduler.client.report [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 116, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 513.004399] env[61728]: DEBUG nova.compute.provider_tree [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 116, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 513.019047] env[61728]: DEBUG nova.scheduler.client.report [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 513.039757] env[61728]: DEBUG nova.scheduler.client.report [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 513.121294] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52db5f0d-0641-75d9-ff41-9a69facf46b8, 'name': SearchDatastore_Task, 'duration_secs': 0.057101} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.121294] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 513.121294] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 513.121294] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 513.121445] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 513.121445] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 513.121445] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dfa796be-b0d0-47be-8485-abc733280e21 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.135652] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 513.135756] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 513.136724] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9be44543-2d7d-4957-9f3b-77459b6e0cd1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.145902] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Waiting for the task: (returnval){ [ 513.145902] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523cb85d-40e6-3a93-32f0-3a531772e7e5" [ 513.145902] env[61728]: _type = "Task" [ 513.145902] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.155131] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523cb85d-40e6-3a93-32f0-3a531772e7e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.193493] env[61728]: DEBUG nova.network.neutron [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Updating instance_info_cache with network_info: [{"id": "89110ed6-c4a4-4f3b-b706-134ce068f320", "address": "fa:16:3e:4c:4f:22", "network": {"id": "4da787ba-7f1e-46fb-9902-d90ec8854a65", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-249168995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "4f587ea8b8fe432aafd3e0ecba054bd0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48937bd1-23dc-413f-b46b-59bf9e709aa0", "external-id": "nsx-vlan-transportzone-160", "segmentation_id": 160, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89110ed6-c4", "ovs_interfaceid": "89110ed6-c4a4-4f3b-b706-134ce068f320", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 513.219684] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463417, 'name': Rename_Task, 'duration_secs': 0.372233} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.220580] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 513.220879] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bdf344c0-b06e-48a9-9d37-c54f661ec098 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.230744] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 513.230744] env[61728]: value = "task-463424" [ 513.230744] env[61728]: _type = "Task" [ 513.230744] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.243919] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463424, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.263235] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463416, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.707035} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.263662] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] d75ab8eb-f15c-4e0f-8376-6d91aa26ab58/d75ab8eb-f15c-4e0f-8376-6d91aa26ab58.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 513.263923] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 513.264266] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9734eb71-b26a-4b9e-b2af-f7c415cdfee3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.274378] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 513.274378] env[61728]: value = "task-463425" [ 513.274378] env[61728]: _type = "Task" [ 513.274378] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.283119] env[61728]: DEBUG nova.network.neutron [req-38335e48-cd09-4e5a-9d9a-a27f6f0d1a35 req-69d6596d-2b22-44ab-9c26-3542168dbb7b service nova] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Updated VIF entry in instance network info cache for port cc02ac93-777e-4bcf-8af5-69f162374735. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 513.283636] env[61728]: DEBUG nova.network.neutron [req-38335e48-cd09-4e5a-9d9a-a27f6f0d1a35 req-69d6596d-2b22-44ab-9c26-3542168dbb7b service nova] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Updating instance_info_cache with network_info: [{"id": "cc02ac93-777e-4bcf-8af5-69f162374735", "address": "fa:16:3e:fc:57:f3", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc02ac93-77", "ovs_interfaceid": "cc02ac93-777e-4bcf-8af5-69f162374735", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 513.296760] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463425, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.319331] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463418, 'name': ReconfigVM_Task, 'duration_secs': 0.500089} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.319694] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Reconfigured VM instance instance-00000002 to attach disk [datastore1] 3cd416c6-9926-45ee-afe3-d3468e4e06c2/3cd416c6-9926-45ee-afe3-d3468e4e06c2.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 513.320493] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36fb900a-484f-428c-b3d4-4106b66b1264 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.331108] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Waiting for the task: (returnval){ [ 513.331108] env[61728]: value = "task-463426" [ 513.331108] env[61728]: _type = "Task" [ 513.331108] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.344299] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463426, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.358110] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463419, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.384842] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d269cb24-b5c0-4881-9160-1dadb5f2261b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.394486] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1974f4d-681e-4ae1-9cf6-eb792f0c8a52 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.433264] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af73e8e-cf99-41c4-a0b7-daf6682b3531 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.442990] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58c38a7-2e6d-4f9f-946c-afd0e48778e5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.460716] env[61728]: DEBUG nova.compute.provider_tree [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 513.660356] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523cb85d-40e6-3a93-32f0-3a531772e7e5, 'name': SearchDatastore_Task, 'duration_secs': 0.054858} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.661198] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-735841ee-0b0c-42af-92d7-0e1163cd26e7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.667172] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Waiting for the task: (returnval){ [ 513.667172] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d1f436-5f65-07ad-8082-9162d55c50cd" [ 513.667172] env[61728]: _type = "Task" [ 513.667172] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.677023] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d1f436-5f65-07ad-8082-9162d55c50cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.697217] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Releasing lock "refresh_cache-26220c0c-f535-4315-b3bb-2e5ac4d2286e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 513.698009] env[61728]: DEBUG nova.compute.manager [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Instance network_info: |[{"id": "89110ed6-c4a4-4f3b-b706-134ce068f320", "address": "fa:16:3e:4c:4f:22", "network": {"id": "4da787ba-7f1e-46fb-9902-d90ec8854a65", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-249168995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "4f587ea8b8fe432aafd3e0ecba054bd0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48937bd1-23dc-413f-b46b-59bf9e709aa0", "external-id": "nsx-vlan-transportzone-160", "segmentation_id": 160, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89110ed6-c4", "ovs_interfaceid": "89110ed6-c4a4-4f3b-b706-134ce068f320", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 513.698515] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4c:4f:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48937bd1-23dc-413f-b46b-59bf9e709aa0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '89110ed6-c4a4-4f3b-b706-134ce068f320', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 513.710895] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Creating folder: Project (4f587ea8b8fe432aafd3e0ecba054bd0). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 513.715859] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-911b5a5d-f041-45c0-af86-ab7cc6cc6a2d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.731192] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Created folder: Project (4f587ea8b8fe432aafd3e0ecba054bd0) in parent group-v121913. [ 513.731497] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Creating folder: Instances. Parent ref: group-v121932. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 513.740638] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c632b5d3-dc23-49bf-9cb7-204d7abab03b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.751213] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463424, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.753124] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Created folder: Instances in parent group-v121932. [ 513.753434] env[61728]: DEBUG oslo.service.loopingcall [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 513.753693] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 513.756876] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f0b9b472-d3f7-47b8-92ac-df89bbf0489f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.786292] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 513.786292] env[61728]: value = "task-463429" [ 513.786292] env[61728]: _type = "Task" [ 513.786292] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.787690] env[61728]: DEBUG oslo_concurrency.lockutils [req-38335e48-cd09-4e5a-9d9a-a27f6f0d1a35 req-69d6596d-2b22-44ab-9c26-3542168dbb7b service nova] Releasing lock "refresh_cache-771341ed-8b8e-470c-9686-82650f5271b4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 513.796179] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463425, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.115223} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.797030] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 513.798201] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894a8bac-a1a4-4fff-9cd9-9f0f624eb943 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.806976] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463429, 'name': CreateVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.830358] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Reconfiguring VM instance instance-00000003 to attach disk [datastore1] d75ab8eb-f15c-4e0f-8376-6d91aa26ab58/d75ab8eb-f15c-4e0f-8376-6d91aa26ab58.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 513.833753] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e544cf72-ac82-49b0-bf7c-4ee779d1a3e0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.865540] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463419, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.631167} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.865940] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463426, 'name': Rename_Task, 'duration_secs': 0.189409} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.867041] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4/771341ed-8b8e-470c-9686-82650f5271b4.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 513.867354] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 513.867604] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 513.867914] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 513.867914] env[61728]: value = "task-463430" [ 513.867914] env[61728]: _type = "Task" [ 513.867914] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.868176] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c81b671d-bd16-44df-8392-cf54519eae96 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.870163] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a320c06-424d-41ee-9ea9-014d37c2d93a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.879124] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 513.879124] env[61728]: value = "task-463431" [ 513.879124] env[61728]: _type = "Task" [ 513.879124] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.884089] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Waiting for the task: (returnval){ [ 513.884089] env[61728]: value = "task-463432" [ 513.884089] env[61728]: _type = "Task" [ 513.884089] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.884089] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463430, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.896870] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463431, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.901452] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463432, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.965793] env[61728]: DEBUG nova.network.neutron [req-e326449b-732f-4373-bd5c-0da2d3e9ae33 req-e509f8a6-2446-41c0-a0f7-349070739fce service nova] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Updated VIF entry in instance network info cache for port e083e0fd-486c-40fb-b54d-c58808ccf0a3. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 513.967638] env[61728]: DEBUG nova.network.neutron [req-e326449b-732f-4373-bd5c-0da2d3e9ae33 req-e509f8a6-2446-41c0-a0f7-349070739fce service nova] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Updating instance_info_cache with network_info: [{"id": "e083e0fd-486c-40fb-b54d-c58808ccf0a3", "address": "fa:16:3e:05:49:1a", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape083e0fd-48", "ovs_interfaceid": "e083e0fd-486c-40fb-b54d-c58808ccf0a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 514.013659] env[61728]: DEBUG nova.scheduler.client.report [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 20 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 514.015089] env[61728]: DEBUG nova.compute.provider_tree [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 20 to 21 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 514.015089] env[61728]: DEBUG nova.compute.provider_tree [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 514.179330] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d1f436-5f65-07ad-8082-9162d55c50cd, 'name': SearchDatastore_Task, 'duration_secs': 0.024053} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 514.179765] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 514.179881] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 08ad3f4a-12a8-48d3-8829-f6744648a95e/08ad3f4a-12a8-48d3-8829-f6744648a95e.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 514.180272] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c3bce063-87f3-4976-bdb4-38819c6a88f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.186707] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Waiting for the task: (returnval){ [ 514.186707] env[61728]: value = "task-463433" [ 514.186707] env[61728]: _type = "Task" [ 514.186707] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 514.194617] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463433, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.239839] env[61728]: DEBUG oslo_vmware.api [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463424, 'name': PowerOnVM_Task, 'duration_secs': 0.555726} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 514.240218] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 514.240481] env[61728]: INFO nova.compute.manager [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Took 15.11 seconds to spawn the instance on the hypervisor. [ 514.240753] env[61728]: DEBUG nova.compute.manager [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 514.241647] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b746a9f-88a2-42b9-8223-453e2f7b17a7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.252399] env[61728]: DEBUG nova.network.neutron [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Successfully updated port: 0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 514.298479] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463429, 'name': CreateVM_Task, 'duration_secs': 0.397453} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 514.298648] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 514.299334] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 514.299524] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 514.299845] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 514.300133] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e019d743-6f98-4491-9d6e-1ede8b0b84fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.306118] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 514.306118] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52191b41-0bfa-56e9-6b93-2bfbaa840fae" [ 514.306118] env[61728]: _type = "Task" [ 514.306118] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 514.315445] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52191b41-0bfa-56e9-6b93-2bfbaa840fae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.382220] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463430, 'name': ReconfigVM_Task, 'duration_secs': 0.407729} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 514.382605] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Reconfigured VM instance instance-00000003 to attach disk [datastore1] d75ab8eb-f15c-4e0f-8376-6d91aa26ab58/d75ab8eb-f15c-4e0f-8376-6d91aa26ab58.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 514.383346] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-63eb8cb5-fd56-49b1-b1b7-73c5c4f7b96a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.395761] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 514.395761] env[61728]: value = "task-463434" [ 514.395761] env[61728]: _type = "Task" [ 514.395761] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 514.400134] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463431, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090491} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 514.404378] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 514.408649] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c751846-3c4c-4f8c-8c8d-0a56ed2a7cda {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.411515] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463432, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.416810] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463434, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.436321] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Reconfiguring VM instance instance-00000004 to attach disk [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4/771341ed-8b8e-470c-9686-82650f5271b4.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 514.438644] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac3dac56-ee85-40a1-8900-bb729550be1d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.464818] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 514.464818] env[61728]: value = "task-463435" [ 514.464818] env[61728]: _type = "Task" [ 514.464818] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 514.474845] env[61728]: DEBUG oslo_concurrency.lockutils [req-e326449b-732f-4373-bd5c-0da2d3e9ae33 req-e509f8a6-2446-41c0-a0f7-349070739fce service nova] Releasing lock "refresh_cache-08ad3f4a-12a8-48d3-8829-f6744648a95e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 514.475299] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463435, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.529800] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.611s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 514.530363] env[61728]: DEBUG nova.compute.manager [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 514.533152] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.537s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.534816] env[61728]: INFO nova.compute.claims [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 514.595994] env[61728]: DEBUG oslo_concurrency.lockutils [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Acquiring lock "be014f77-53ca-42a5-9255-2ba950cca314" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.596207] env[61728]: DEBUG oslo_concurrency.lockutils [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Lock "be014f77-53ca-42a5-9255-2ba950cca314" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.701455] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463433, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.756428] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "refresh_cache-a3e9af38-5115-4337-a20e-5d6c6ec655fe" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 514.756710] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "refresh_cache-a3e9af38-5115-4337-a20e-5d6c6ec655fe" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 514.756919] env[61728]: DEBUG nova.network.neutron [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 514.764438] env[61728]: INFO nova.compute.manager [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Took 19.88 seconds to build instance. [ 514.815634] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52191b41-0bfa-56e9-6b93-2bfbaa840fae, 'name': SearchDatastore_Task, 'duration_secs': 0.015642} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 514.815975] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 514.816321] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 514.816703] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 514.816703] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 514.816986] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 514.817446] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae7d9f2a-a0f9-4692-b816-8ea6cf7456a2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.834499] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 514.834779] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 514.835438] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ecec30e-4034-4526-ae87-385098d565ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.841239] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 514.841239] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525547d7-e418-ac5b-0594-c8fdaf4fbad5" [ 514.841239] env[61728]: _type = "Task" [ 514.841239] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 514.849258] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525547d7-e418-ac5b-0594-c8fdaf4fbad5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.904448] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463432, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.913986] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463434, 'name': Rename_Task, 'duration_secs': 0.169277} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 514.913986] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 514.914538] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fadab398-03b7-4ff6-95b8-e6bdb50e35f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.921317] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 514.921317] env[61728]: value = "task-463436" [ 514.921317] env[61728]: _type = "Task" [ 514.921317] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 514.931378] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463436, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.978375] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463435, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.042069] env[61728]: DEBUG nova.compute.utils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 515.045970] env[61728]: DEBUG nova.compute.manager [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 515.046194] env[61728]: DEBUG nova.network.neutron [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 515.177040] env[61728]: DEBUG nova.compute.manager [req-46ed5da1-c1b3-4582-a71f-1cea16f19e66 req-c4c7b452-9a5b-4ac7-88b3-e116b71bf604 service nova] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Received event network-vif-plugged-89110ed6-c4a4-4f3b-b706-134ce068f320 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 515.177270] env[61728]: DEBUG oslo_concurrency.lockutils [req-46ed5da1-c1b3-4582-a71f-1cea16f19e66 req-c4c7b452-9a5b-4ac7-88b3-e116b71bf604 service nova] Acquiring lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.178017] env[61728]: DEBUG oslo_concurrency.lockutils [req-46ed5da1-c1b3-4582-a71f-1cea16f19e66 req-c4c7b452-9a5b-4ac7-88b3-e116b71bf604 service nova] Lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.178017] env[61728]: DEBUG oslo_concurrency.lockutils [req-46ed5da1-c1b3-4582-a71f-1cea16f19e66 req-c4c7b452-9a5b-4ac7-88b3-e116b71bf604 service nova] Lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 515.178017] env[61728]: DEBUG nova.compute.manager [req-46ed5da1-c1b3-4582-a71f-1cea16f19e66 req-c4c7b452-9a5b-4ac7-88b3-e116b71bf604 service nova] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] No waiting events found dispatching network-vif-plugged-89110ed6-c4a4-4f3b-b706-134ce068f320 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 515.182317] env[61728]: WARNING nova.compute.manager [req-46ed5da1-c1b3-4582-a71f-1cea16f19e66 req-c4c7b452-9a5b-4ac7-88b3-e116b71bf604 service nova] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Received unexpected event network-vif-plugged-89110ed6-c4a4-4f3b-b706-134ce068f320 for instance with vm_state building and task_state spawning. [ 515.182317] env[61728]: DEBUG nova.compute.manager [req-46ed5da1-c1b3-4582-a71f-1cea16f19e66 req-c4c7b452-9a5b-4ac7-88b3-e116b71bf604 service nova] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Received event network-changed-89110ed6-c4a4-4f3b-b706-134ce068f320 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 515.182317] env[61728]: DEBUG nova.compute.manager [req-46ed5da1-c1b3-4582-a71f-1cea16f19e66 req-c4c7b452-9a5b-4ac7-88b3-e116b71bf604 service nova] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Refreshing instance network info cache due to event network-changed-89110ed6-c4a4-4f3b-b706-134ce068f320. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 515.182317] env[61728]: DEBUG oslo_concurrency.lockutils [req-46ed5da1-c1b3-4582-a71f-1cea16f19e66 req-c4c7b452-9a5b-4ac7-88b3-e116b71bf604 service nova] Acquiring lock "refresh_cache-26220c0c-f535-4315-b3bb-2e5ac4d2286e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 515.182317] env[61728]: DEBUG oslo_concurrency.lockutils [req-46ed5da1-c1b3-4582-a71f-1cea16f19e66 req-c4c7b452-9a5b-4ac7-88b3-e116b71bf604 service nova] Acquired lock "refresh_cache-26220c0c-f535-4315-b3bb-2e5ac4d2286e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 515.182536] env[61728]: DEBUG nova.network.neutron [req-46ed5da1-c1b3-4582-a71f-1cea16f19e66 req-c4c7b452-9a5b-4ac7-88b3-e116b71bf604 service nova] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Refreshing network info cache for port 89110ed6-c4a4-4f3b-b706-134ce068f320 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 515.186120] env[61728]: DEBUG nova.policy [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '38f64fd31712472a841f82c620d9340b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f5329f59d8bf43b2a855c4259387fccd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 515.199781] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463433, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.566113} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 515.201603] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 08ad3f4a-12a8-48d3-8829-f6744648a95e/08ad3f4a-12a8-48d3-8829-f6744648a95e.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 515.201944] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 515.202603] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-387b2c9e-3ba7-49f8-8d8c-bcf884236d17 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.212067] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Waiting for the task: (returnval){ [ 515.212067] env[61728]: value = "task-463437" [ 515.212067] env[61728]: _type = "Task" [ 515.212067] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 515.224080] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463437, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.271025] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb40d777-1c5b-4c59-8650-fb9c697f31b8 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.391s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 515.325865] env[61728]: DEBUG nova.network.neutron [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 515.354820] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525547d7-e418-ac5b-0594-c8fdaf4fbad5, 'name': SearchDatastore_Task, 'duration_secs': 0.020449} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 515.355618] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f5621c4-326d-4fd1-a302-75d31953f643 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.361866] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 515.361866] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5267b4fa-a5fa-4fa4-ec25-c2b72e64ec82" [ 515.361866] env[61728]: _type = "Task" [ 515.361866] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 515.376568] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5267b4fa-a5fa-4fa4-ec25-c2b72e64ec82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.404730] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463432, 'name': PowerOnVM_Task} progress is 91%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.440027] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463436, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.484677] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463435, 'name': ReconfigVM_Task, 'duration_secs': 0.724081} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 515.484677] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Reconfigured VM instance instance-00000004 to attach disk [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4/771341ed-8b8e-470c-9686-82650f5271b4.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 515.484677] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e62fe1b-8c0f-4cad-856e-4d980cbbefaf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.490223] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 515.490223] env[61728]: value = "task-463438" [ 515.490223] env[61728]: _type = "Task" [ 515.490223] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 515.500176] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463438, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.544866] env[61728]: DEBUG nova.compute.manager [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 515.724535] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463437, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093838} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 515.724535] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 515.724956] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b891f2-5921-4d1b-afdb-23d9e14677ce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.751903] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Reconfiguring VM instance instance-00000005 to attach disk [datastore1] 08ad3f4a-12a8-48d3-8829-f6744648a95e/08ad3f4a-12a8-48d3-8829-f6744648a95e.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 515.757449] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8310f49-aed5-47c6-95fe-33ff59f3e32b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.776962] env[61728]: DEBUG nova.compute.manager [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.786456] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Waiting for the task: (returnval){ [ 515.786456] env[61728]: value = "task-463439" [ 515.786456] env[61728]: _type = "Task" [ 515.786456] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 515.796232] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463439, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.891792] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5267b4fa-a5fa-4fa4-ec25-c2b72e64ec82, 'name': SearchDatastore_Task, 'duration_secs': 0.014034} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 515.892728] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 515.893170] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 26220c0c-f535-4315-b3bb-2e5ac4d2286e/26220c0c-f535-4315-b3bb-2e5ac4d2286e.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 515.893516] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-63acdcaf-3f03-4840-86cc-e6215676160b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.909313] env[61728]: DEBUG oslo_vmware.api [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463432, 'name': PowerOnVM_Task, 'duration_secs': 1.556726} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 515.914675] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 515.915069] env[61728]: INFO nova.compute.manager [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Took 15.11 seconds to spawn the instance on the hypervisor. [ 515.915276] env[61728]: DEBUG nova.compute.manager [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 515.915601] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 515.915601] env[61728]: value = "task-463440" [ 515.915601] env[61728]: _type = "Task" [ 515.915601] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 515.916544] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc78664-c9f2-4be8-9bea-35a95fe47f11 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.937732] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463440, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.943052] env[61728]: DEBUG nova.network.neutron [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Updating instance_info_cache with network_info: [{"id": "0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4", "address": "fa:16:3e:4e:88:29", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a12ba5f-fa", "ovs_interfaceid": "0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 515.951938] env[61728]: DEBUG nova.compute.manager [req-db90081a-b93d-4b1a-add2-79ad73ddcbce req-6049219e-d02d-4f9b-afcf-15e8a53c829b service nova] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Received event network-vif-plugged-0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 515.951938] env[61728]: DEBUG oslo_concurrency.lockutils [req-db90081a-b93d-4b1a-add2-79ad73ddcbce req-6049219e-d02d-4f9b-afcf-15e8a53c829b service nova] Acquiring lock "a3e9af38-5115-4337-a20e-5d6c6ec655fe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.952129] env[61728]: DEBUG oslo_concurrency.lockutils [req-db90081a-b93d-4b1a-add2-79ad73ddcbce req-6049219e-d02d-4f9b-afcf-15e8a53c829b service nova] Lock "a3e9af38-5115-4337-a20e-5d6c6ec655fe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.952272] env[61728]: DEBUG oslo_concurrency.lockutils [req-db90081a-b93d-4b1a-add2-79ad73ddcbce req-6049219e-d02d-4f9b-afcf-15e8a53c829b service nova] Lock "a3e9af38-5115-4337-a20e-5d6c6ec655fe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 515.952430] env[61728]: DEBUG nova.compute.manager [req-db90081a-b93d-4b1a-add2-79ad73ddcbce req-6049219e-d02d-4f9b-afcf-15e8a53c829b service nova] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] No waiting events found dispatching network-vif-plugged-0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 515.952616] env[61728]: WARNING nova.compute.manager [req-db90081a-b93d-4b1a-add2-79ad73ddcbce req-6049219e-d02d-4f9b-afcf-15e8a53c829b service nova] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Received unexpected event network-vif-plugged-0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4 for instance with vm_state building and task_state spawning. [ 515.952789] env[61728]: DEBUG nova.compute.manager [req-db90081a-b93d-4b1a-add2-79ad73ddcbce req-6049219e-d02d-4f9b-afcf-15e8a53c829b service nova] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Received event network-changed-0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 515.952937] env[61728]: DEBUG nova.compute.manager [req-db90081a-b93d-4b1a-add2-79ad73ddcbce req-6049219e-d02d-4f9b-afcf-15e8a53c829b service nova] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Refreshing instance network info cache due to event network-changed-0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 515.953331] env[61728]: DEBUG oslo_concurrency.lockutils [req-db90081a-b93d-4b1a-add2-79ad73ddcbce req-6049219e-d02d-4f9b-afcf-15e8a53c829b service nova] Acquiring lock "refresh_cache-a3e9af38-5115-4337-a20e-5d6c6ec655fe" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 515.956795] env[61728]: DEBUG oslo_vmware.api [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463436, 'name': PowerOnVM_Task, 'duration_secs': 0.754293} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 515.956795] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 515.956795] env[61728]: INFO nova.compute.manager [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Took 12.87 seconds to spawn the instance on the hypervisor. [ 515.956795] env[61728]: DEBUG nova.compute.manager [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 515.957278] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-768ce44d-ac30-472e-8212-b2cdb9abb2e8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.001971] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463438, 'name': Rename_Task, 'duration_secs': 0.21872} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 516.002257] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 516.002510] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb119e7c-b2fd-4e53-b7ac-af76f0ae9ce9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.014708] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 516.014708] env[61728]: value = "task-463441" [ 516.014708] env[61728]: _type = "Task" [ 516.014708] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.032600] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463441, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.048718] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4bf524d-3574-4b64-8bd7-60bec092fbf6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.060135] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68fd68a9-05e2-4d71-ba43-594ad38c289e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.096079] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37a3072-3dd9-485d-8415-b46c4f111d17 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.105882] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be905313-d7cb-431a-a086-aaef75b091ce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.122168] env[61728]: DEBUG nova.compute.provider_tree [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 516.127190] env[61728]: DEBUG nova.network.neutron [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Successfully created port: 86d2fa33-d261-42a1-b49d-1fd478f8311b {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 516.294166] env[61728]: DEBUG nova.network.neutron [req-46ed5da1-c1b3-4582-a71f-1cea16f19e66 req-c4c7b452-9a5b-4ac7-88b3-e116b71bf604 service nova] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Updated VIF entry in instance network info cache for port 89110ed6-c4a4-4f3b-b706-134ce068f320. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 516.294566] env[61728]: DEBUG nova.network.neutron [req-46ed5da1-c1b3-4582-a71f-1cea16f19e66 req-c4c7b452-9a5b-4ac7-88b3-e116b71bf604 service nova] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Updating instance_info_cache with network_info: [{"id": "89110ed6-c4a4-4f3b-b706-134ce068f320", "address": "fa:16:3e:4c:4f:22", "network": {"id": "4da787ba-7f1e-46fb-9902-d90ec8854a65", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-249168995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "4f587ea8b8fe432aafd3e0ecba054bd0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48937bd1-23dc-413f-b46b-59bf9e709aa0", "external-id": "nsx-vlan-transportzone-160", "segmentation_id": 160, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89110ed6-c4", "ovs_interfaceid": "89110ed6-c4a4-4f3b-b706-134ce068f320", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 516.302076] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463439, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.311319] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.430723] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463440, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.465271] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "refresh_cache-a3e9af38-5115-4337-a20e-5d6c6ec655fe" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 516.465715] env[61728]: DEBUG nova.compute.manager [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Instance network_info: |[{"id": "0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4", "address": "fa:16:3e:4e:88:29", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a12ba5f-fa", "ovs_interfaceid": "0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 516.467933] env[61728]: DEBUG oslo_concurrency.lockutils [req-db90081a-b93d-4b1a-add2-79ad73ddcbce req-6049219e-d02d-4f9b-afcf-15e8a53c829b service nova] Acquired lock "refresh_cache-a3e9af38-5115-4337-a20e-5d6c6ec655fe" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 516.468144] env[61728]: DEBUG nova.network.neutron [req-db90081a-b93d-4b1a-add2-79ad73ddcbce req-6049219e-d02d-4f9b-afcf-15e8a53c829b service nova] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Refreshing network info cache for port 0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 516.469332] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:88:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f65996a3-f865-4492-9377-cd14ec8b3aae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 516.480020] env[61728]: DEBUG oslo.service.loopingcall [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 516.480652] env[61728]: INFO nova.compute.manager [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Took 21.13 seconds to build instance. [ 516.486232] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 516.487357] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2df83c20-f0ab-4096-8bab-7685ba895d5b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.506342] env[61728]: INFO nova.compute.manager [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Took 20.50 seconds to build instance. [ 516.511997] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 516.511997] env[61728]: value = "task-463443" [ 516.511997] env[61728]: _type = "Task" [ 516.511997] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.525172] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463443, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.529735] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463441, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.558296] env[61728]: DEBUG nova.compute.manager [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 516.602207] env[61728]: DEBUG nova.virt.hardware [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 516.602625] env[61728]: DEBUG nova.virt.hardware [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 516.602832] env[61728]: DEBUG nova.virt.hardware [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 516.603372] env[61728]: DEBUG nova.virt.hardware [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 516.603372] env[61728]: DEBUG nova.virt.hardware [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 516.603372] env[61728]: DEBUG nova.virt.hardware [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 516.603528] env[61728]: DEBUG nova.virt.hardware [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 516.603880] env[61728]: DEBUG nova.virt.hardware [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 516.604082] env[61728]: DEBUG nova.virt.hardware [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 516.604874] env[61728]: DEBUG nova.virt.hardware [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 516.604874] env[61728]: DEBUG nova.virt.hardware [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 516.605657] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4147f1-8f20-4fd4-8fb1-1bd8ce9aa675 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.614638] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb0c346-45d0-455c-9740-ba09f404b474 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.632082] env[61728]: DEBUG nova.scheduler.client.report [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 516.800172] env[61728]: DEBUG oslo_concurrency.lockutils [req-46ed5da1-c1b3-4582-a71f-1cea16f19e66 req-c4c7b452-9a5b-4ac7-88b3-e116b71bf604 service nova] Releasing lock "refresh_cache-26220c0c-f535-4315-b3bb-2e5ac4d2286e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 516.800811] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463439, 'name': ReconfigVM_Task, 'duration_secs': 0.834707} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 516.801138] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Reconfigured VM instance instance-00000005 to attach disk [datastore1] 08ad3f4a-12a8-48d3-8829-f6744648a95e/08ad3f4a-12a8-48d3-8829-f6744648a95e.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 516.802026] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7c82f575-f03e-4979-a1da-83133b37c573 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.809173] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Waiting for the task: (returnval){ [ 516.809173] env[61728]: value = "task-463444" [ 516.809173] env[61728]: _type = "Task" [ 516.809173] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.818687] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463444, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.932033] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463440, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.827723} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 516.936910] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 26220c0c-f535-4315-b3bb-2e5ac4d2286e/26220c0c-f535-4315-b3bb-2e5ac4d2286e.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 516.937248] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 516.938353] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-72369365-a261-4d5f-9189-026dbff63f09 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.945886] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 516.945886] env[61728]: value = "task-463445" [ 516.945886] env[61728]: _type = "Task" [ 516.945886] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.958228] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463445, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.987437] env[61728]: DEBUG oslo_concurrency.lockutils [None req-82504c7c-8799-4849-a49f-ac3ffb4af602 tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Lock "3cd416c6-9926-45ee-afe3-d3468e4e06c2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.655s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.008941] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4817c722-6bf5-4e54-8e51-8fc08503a837 tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Lock "d75ab8eb-f15c-4e0f-8376-6d91aa26ab58" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.017s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.031954] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463443, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 517.035506] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463441, 'name': PowerOnVM_Task} progress is 79%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 517.137696] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.604s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.139681] env[61728]: DEBUG nova.compute.manager [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 517.144746] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.869s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.322634] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463444, 'name': Rename_Task, 'duration_secs': 0.403765} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 517.322923] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 517.323174] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c1b4be68-3249-4124-b687-056a699617f9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.331816] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Waiting for the task: (returnval){ [ 517.331816] env[61728]: value = "task-463446" [ 517.331816] env[61728]: _type = "Task" [ 517.331816] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 517.340297] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463446, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 517.459717] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463445, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.101253} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 517.460388] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 517.461230] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d76651-3d56-483e-a8c8-a129e2bd351c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.486338] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Reconfiguring VM instance instance-00000006 to attach disk [datastore1] 26220c0c-f535-4315-b3bb-2e5ac4d2286e/26220c0c-f535-4315-b3bb-2e5ac4d2286e.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 517.486615] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bcae4122-1441-4467-9c9f-a93fd327ecac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.502079] env[61728]: DEBUG nova.compute.manager [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.510152] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 517.510152] env[61728]: value = "task-463447" [ 517.510152] env[61728]: _type = "Task" [ 517.510152] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 517.513887] env[61728]: DEBUG nova.compute.manager [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.526555] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463447, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 517.534656] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463443, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 517.545757] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463441, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 517.648498] env[61728]: DEBUG nova.compute.utils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 517.659066] env[61728]: DEBUG nova.network.neutron [req-db90081a-b93d-4b1a-add2-79ad73ddcbce req-6049219e-d02d-4f9b-afcf-15e8a53c829b service nova] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Updated VIF entry in instance network info cache for port 0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 517.659413] env[61728]: DEBUG nova.network.neutron [req-db90081a-b93d-4b1a-add2-79ad73ddcbce req-6049219e-d02d-4f9b-afcf-15e8a53c829b service nova] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Updating instance_info_cache with network_info: [{"id": "0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4", "address": "fa:16:3e:4e:88:29", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a12ba5f-fa", "ovs_interfaceid": "0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 517.660893] env[61728]: DEBUG nova.compute.manager [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 517.660994] env[61728]: DEBUG nova.network.neutron [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 517.709857] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "db8e373d-e06c-43f0-a79c-2c1aaf65a01a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.709857] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "db8e373d-e06c-43f0-a79c-2c1aaf65a01a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.740529] env[61728]: DEBUG nova.policy [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd32b5118a94042f797dea1fdf6bd05b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08367aaceba548fe93faaedf6371817d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 517.847433] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463446, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.047571] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.058259] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463447, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.058629] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463443, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.069387] env[61728]: DEBUG oslo_vmware.api [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463441, 'name': PowerOnVM_Task, 'duration_secs': 1.562668} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 518.069668] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.069967] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 518.070178] env[61728]: INFO nova.compute.manager [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Took 12.78 seconds to spawn the instance on the hypervisor. [ 518.070351] env[61728]: DEBUG nova.compute.manager [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 518.071162] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d1d6d2e-d366-41fc-89cd-dda67cc1e7c4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.097950] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "c8f1f4ae-ab08-4a03-a3fd-94014509e7e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.098243] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "c8f1f4ae-ab08-4a03-a3fd-94014509e7e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.161351] env[61728]: DEBUG nova.compute.manager [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 518.173255] env[61728]: DEBUG oslo_concurrency.lockutils [req-db90081a-b93d-4b1a-add2-79ad73ddcbce req-6049219e-d02d-4f9b-afcf-15e8a53c829b service nova] Releasing lock "refresh_cache-a3e9af38-5115-4337-a20e-5d6c6ec655fe" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 518.202169] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance d75ab8eb-f15c-4e0f-8376-6d91aa26ab58 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 518.202299] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 518.202429] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 3cd416c6-9926-45ee-afe3-d3468e4e06c2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 518.202595] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 771341ed-8b8e-470c-9686-82650f5271b4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 518.202707] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 08ad3f4a-12a8-48d3-8829-f6744648a95e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 518.202908] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 26220c0c-f535-4315-b3bb-2e5ac4d2286e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 518.202955] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance a3e9af38-5115-4337-a20e-5d6c6ec655fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 518.203048] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 518.203155] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 4247e356-ef10-41b3-bb77-106103134d59 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 518.345502] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463446, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.384778] env[61728]: DEBUG nova.network.neutron [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Successfully created port: 590df666-2609-450c-b0e2-e5df320ae98b {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 518.532521] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463443, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.536553] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463447, 'name': ReconfigVM_Task, 'duration_secs': 0.818874} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 518.536872] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Reconfigured VM instance instance-00000006 to attach disk [datastore1] 26220c0c-f535-4315-b3bb-2e5ac4d2286e/26220c0c-f535-4315-b3bb-2e5ac4d2286e.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 518.537531] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-97fe5b5e-9618-476b-9089-5d4147a28491 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.544165] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 518.544165] env[61728]: value = "task-463449" [ 518.544165] env[61728]: _type = "Task" [ 518.544165] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 518.553394] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463449, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.592593] env[61728]: INFO nova.compute.manager [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Took 17.58 seconds to build instance. [ 518.707831] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance b86bf217-d1f8-4d37-aa87-3a2d27d70420 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 518.847358] env[61728]: DEBUG oslo_vmware.api [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463446, 'name': PowerOnVM_Task, 'duration_secs': 1.489159} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 518.847818] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 518.848043] env[61728]: INFO nova.compute.manager [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Took 10.66 seconds to spawn the instance on the hypervisor. [ 518.848255] env[61728]: DEBUG nova.compute.manager [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 518.849073] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-906a4d52-f182-4377-be36-4f3ff2c556ea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.920481] env[61728]: DEBUG nova.network.neutron [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Successfully updated port: 86d2fa33-d261-42a1-b49d-1fd478f8311b {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 519.026162] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463443, 'name': CreateVM_Task, 'duration_secs': 2.021568} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 519.026162] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 519.027825] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 519.028127] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 519.028924] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 519.029311] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8f5dcaa-db4b-4679-875f-6625b5ac94bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.035183] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 519.035183] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fa9447-9add-88f3-dd9d-d365cfb0bcdf" [ 519.035183] env[61728]: _type = "Task" [ 519.035183] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.045913] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fa9447-9add-88f3-dd9d-d365cfb0bcdf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.053629] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463449, 'name': Rename_Task, 'duration_secs': 0.176295} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 519.054220] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 519.055167] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d154105-145c-4901-9dca-2ad2d6b565f1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.062818] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 519.062818] env[61728]: value = "task-463450" [ 519.062818] env[61728]: _type = "Task" [ 519.062818] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.073427] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463450, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.094312] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c63821b3-e6cd-459f-9431-fa4572bac629 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "771341ed-8b8e-470c-9686-82650f5271b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.097s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.185990] env[61728]: DEBUG nova.compute.manager [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 519.218105] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance adb58292-62cd-4142-ad66-db014bcbdc49 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 519.239184] env[61728]: DEBUG nova.virt.hardware [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 519.239436] env[61728]: DEBUG nova.virt.hardware [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 519.239640] env[61728]: DEBUG nova.virt.hardware [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 519.239832] env[61728]: DEBUG nova.virt.hardware [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 519.239992] env[61728]: DEBUG nova.virt.hardware [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 519.240264] env[61728]: DEBUG nova.virt.hardware [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 519.240485] env[61728]: DEBUG nova.virt.hardware [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 519.240643] env[61728]: DEBUG nova.virt.hardware [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 519.241488] env[61728]: DEBUG nova.virt.hardware [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 519.241488] env[61728]: DEBUG nova.virt.hardware [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 519.242260] env[61728]: DEBUG nova.virt.hardware [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 519.244980] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34514228-e3c6-49e5-96fb-044a7e8a9ac9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.255316] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62adc898-b4e0-4264-96b7-9588c65ca05e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.371300] env[61728]: INFO nova.compute.manager [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Took 15.62 seconds to build instance. [ 519.426577] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "refresh_cache-ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 519.426731] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "refresh_cache-ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 519.426898] env[61728]: DEBUG nova.network.neutron [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 519.454046] env[61728]: DEBUG nova.compute.manager [None req-571100e5-3025-4cfd-b895-20af0009d2c6 tempest-ServerDiagnosticsTest-1784047041 tempest-ServerDiagnosticsTest-1784047041-project-admin] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 519.456324] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405472a6-32ad-4c45-8e08-ba64a9bedb28 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.464879] env[61728]: INFO nova.compute.manager [None req-571100e5-3025-4cfd-b895-20af0009d2c6 tempest-ServerDiagnosticsTest-1784047041 tempest-ServerDiagnosticsTest-1784047041-project-admin] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Retrieving diagnostics [ 519.465897] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b340912-396e-4cf9-946e-fd1989ae3ac8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.549200] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fa9447-9add-88f3-dd9d-d365cfb0bcdf, 'name': SearchDatastore_Task, 'duration_secs': 0.040579} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 519.549200] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 519.549200] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 519.549367] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 519.549535] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 519.549728] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 519.550058] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9855aa1-8eb2-4807-abb2-cd595b2a8b3c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.560169] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 519.560221] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 519.561044] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c16aac13-ec8a-403c-b660-858ede0fa306 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.573055] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 519.573055] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52013b64-b30d-edcc-a258-0c9a9e5f1ae7" [ 519.573055] env[61728]: _type = "Task" [ 519.573055] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.580226] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463450, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.589865] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52013b64-b30d-edcc-a258-0c9a9e5f1ae7, 'name': SearchDatastore_Task, 'duration_secs': 0.011841} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 519.592751] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14d9d7ab-7d61-42c2-9984-3b9918c5d72e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.600891] env[61728]: DEBUG nova.compute.manager [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 519.609252] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 519.609252] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f89ca4-e1f4-ffe6-d85a-074aa2e72edb" [ 519.609252] env[61728]: _type = "Task" [ 519.609252] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.621850] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f89ca4-e1f4-ffe6-d85a-074aa2e72edb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.729248] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance e56a59a6-fc0f-45ca-810a-76b18bae173a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 519.731049] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Acquiring lock "d75ab8eb-f15c-4e0f-8376-6d91aa26ab58" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.734664] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Lock "d75ab8eb-f15c-4e0f-8376-6d91aa26ab58" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.734664] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Acquiring lock "d75ab8eb-f15c-4e0f-8376-6d91aa26ab58-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.734664] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Lock "d75ab8eb-f15c-4e0f-8376-6d91aa26ab58-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.734664] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Lock "d75ab8eb-f15c-4e0f-8376-6d91aa26ab58-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.734933] env[61728]: INFO nova.compute.manager [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Terminating instance [ 519.738171] env[61728]: DEBUG nova.compute.manager [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 519.738350] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 519.741022] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a02323-9432-44c6-aac7-cb3c1bad95a2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.748150] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 519.748385] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a6ae45d-dc2c-48d9-afb4-03d4af892e37 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.756144] env[61728]: DEBUG oslo_vmware.api [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Waiting for the task: (returnval){ [ 519.756144] env[61728]: value = "task-463451" [ 519.756144] env[61728]: _type = "Task" [ 519.756144] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.765779] env[61728]: DEBUG oslo_vmware.api [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Task: {'id': task-463451, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.873332] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa42974-e32b-4fd6-9ca2-46434e2bc566 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Lock "08ad3f4a-12a8-48d3-8829-f6744648a95e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.131s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.884615] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Acquiring lock "2c76f45a-b87c-4c3b-9bf9-ba5d67265437" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.884860] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Lock "2c76f45a-b87c-4c3b-9bf9-ba5d67265437" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.964750] env[61728]: DEBUG nova.network.neutron [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 520.077388] env[61728]: DEBUG oslo_vmware.api [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463450, 'name': PowerOnVM_Task, 'duration_secs': 0.829791} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 520.077388] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 520.077388] env[61728]: INFO nova.compute.manager [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Took 9.49 seconds to spawn the instance on the hypervisor. [ 520.077388] env[61728]: DEBUG nova.compute.manager [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 520.077616] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf468fc-5431-41be-973c-efb2c2679117 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.126172] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f89ca4-e1f4-ffe6-d85a-074aa2e72edb, 'name': SearchDatastore_Task, 'duration_secs': 0.010473} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 520.126172] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 520.126172] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a3e9af38-5115-4337-a20e-5d6c6ec655fe/a3e9af38-5115-4337-a20e-5d6c6ec655fe.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 520.126172] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-60ca3655-7129-4d2d-a5f9-2a2696ddbde8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.135444] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 520.135444] env[61728]: value = "task-463452" [ 520.135444] env[61728]: _type = "Task" [ 520.135444] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.138227] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.151388] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463452, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.219601] env[61728]: DEBUG nova.network.neutron [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Successfully updated port: 590df666-2609-450c-b0e2-e5df320ae98b {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 520.226734] env[61728]: DEBUG nova.compute.manager [req-b09c3710-a137-44cd-ac94-3e2fcec05fdf req-23d91bb7-0ae0-4aee-b1bc-229f9ef83f2b service nova] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Received event network-vif-plugged-86d2fa33-d261-42a1-b49d-1fd478f8311b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 520.227737] env[61728]: DEBUG oslo_concurrency.lockutils [req-b09c3710-a137-44cd-ac94-3e2fcec05fdf req-23d91bb7-0ae0-4aee-b1bc-229f9ef83f2b service nova] Acquiring lock "ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.227737] env[61728]: DEBUG oslo_concurrency.lockutils [req-b09c3710-a137-44cd-ac94-3e2fcec05fdf req-23d91bb7-0ae0-4aee-b1bc-229f9ef83f2b service nova] Lock "ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.227737] env[61728]: DEBUG oslo_concurrency.lockutils [req-b09c3710-a137-44cd-ac94-3e2fcec05fdf req-23d91bb7-0ae0-4aee-b1bc-229f9ef83f2b service nova] Lock "ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.227737] env[61728]: DEBUG nova.compute.manager [req-b09c3710-a137-44cd-ac94-3e2fcec05fdf req-23d91bb7-0ae0-4aee-b1bc-229f9ef83f2b service nova] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] No waiting events found dispatching network-vif-plugged-86d2fa33-d261-42a1-b49d-1fd478f8311b {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 520.227737] env[61728]: WARNING nova.compute.manager [req-b09c3710-a137-44cd-ac94-3e2fcec05fdf req-23d91bb7-0ae0-4aee-b1bc-229f9ef83f2b service nova] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Received unexpected event network-vif-plugged-86d2fa33-d261-42a1-b49d-1fd478f8311b for instance with vm_state building and task_state spawning. [ 520.235071] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 3c4f7d53-6ba6-4e2c-b324-e651640b304b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 520.270145] env[61728]: DEBUG oslo_vmware.api [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Task: {'id': task-463451, 'name': PowerOffVM_Task, 'duration_secs': 0.246904} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 520.270468] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 520.270674] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 520.270957] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0ec3bcf5-6249-4ff5-bcf7-1b6c11217acf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.334731] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 520.334966] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 520.335245] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Deleting the datastore file [datastore1] d75ab8eb-f15c-4e0f-8376-6d91aa26ab58 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 520.335513] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bdbf7408-4f7a-4db9-bbe3-e3d284aa836a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.341874] env[61728]: DEBUG oslo_vmware.api [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Waiting for the task: (returnval){ [ 520.341874] env[61728]: value = "task-463455" [ 520.341874] env[61728]: _type = "Task" [ 520.341874] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.349943] env[61728]: DEBUG nova.compute.manager [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Stashing vm_state: active {{(pid=61728) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 520.355394] env[61728]: DEBUG oslo_vmware.api [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Task: {'id': task-463455, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.378038] env[61728]: DEBUG nova.compute.manager [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 520.407540] env[61728]: DEBUG nova.network.neutron [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Updating instance_info_cache with network_info: [{"id": "86d2fa33-d261-42a1-b49d-1fd478f8311b", "address": "fa:16:3e:f5:2d:48", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86d2fa33-d2", "ovs_interfaceid": "86d2fa33-d261-42a1-b49d-1fd478f8311b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.605188] env[61728]: INFO nova.compute.manager [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Took 16.07 seconds to build instance. [ 520.648772] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463452, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.727049] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "refresh_cache-4247e356-ef10-41b3-bb77-106103134d59" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 520.727049] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "refresh_cache-4247e356-ef10-41b3-bb77-106103134d59" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 520.727049] env[61728]: DEBUG nova.network.neutron [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 520.738718] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 620bb472-c36e-4c56-acdf-42e02614856b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 520.854285] env[61728]: DEBUG oslo_vmware.api [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Task: {'id': task-463455, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.881219] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.900331] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.910227] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "refresh_cache-ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 520.910562] env[61728]: DEBUG nova.compute.manager [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Instance network_info: |[{"id": "86d2fa33-d261-42a1-b49d-1fd478f8311b", "address": "fa:16:3e:f5:2d:48", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86d2fa33-d2", "ovs_interfaceid": "86d2fa33-d261-42a1-b49d-1fd478f8311b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 520.911053] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f5:2d:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c492f5cc-7ae0-4cab-823c-0d5dd8c60b26', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '86d2fa33-d261-42a1-b49d-1fd478f8311b', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 520.922671] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Creating folder: Project (f5329f59d8bf43b2a855c4259387fccd). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 520.922974] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d82a9c47-3388-43c0-8cbd-d96ff6f73957 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.934740] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Created folder: Project (f5329f59d8bf43b2a855c4259387fccd) in parent group-v121913. [ 520.934740] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Creating folder: Instances. Parent ref: group-v121939. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 520.934831] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a640809f-0a27-4139-bac6-365cd0963bd5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.944269] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Created folder: Instances in parent group-v121939. [ 520.944644] env[61728]: DEBUG oslo.service.loopingcall [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 520.944884] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 520.945108] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2adb5e9b-4393-4b5b-9ea5-fe1dd1705d3a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.964714] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 520.964714] env[61728]: value = "task-463461" [ 520.964714] env[61728]: _type = "Task" [ 520.964714] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.972812] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463461, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.108176] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d224f82-035b-4eef-8cb4-c36ed3e41492 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.577s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 521.154050] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463452, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.621226} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.154597] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a3e9af38-5115-4337-a20e-5d6c6ec655fe/a3e9af38-5115-4337-a20e-5d6c6ec655fe.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 521.154597] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 521.154819] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c6aca4fb-275a-4f8f-b01a-35865f01ee08 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.161224] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 521.161224] env[61728]: value = "task-463462" [ 521.161224] env[61728]: _type = "Task" [ 521.161224] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.169295] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463462, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.241274] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 8c4b7227-1a01-4400-beb8-2df27e17e329 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 521.274078] env[61728]: DEBUG nova.network.neutron [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 521.353862] env[61728]: DEBUG oslo_vmware.api [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Task: {'id': task-463455, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.539886} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.353862] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 521.353862] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 521.354389] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 521.354389] env[61728]: INFO nova.compute.manager [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Took 1.62 seconds to destroy the instance on the hypervisor. [ 521.354389] env[61728]: DEBUG oslo.service.loopingcall [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 521.354534] env[61728]: DEBUG nova.compute.manager [-] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 521.354625] env[61728]: DEBUG nova.network.neutron [-] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 521.487292] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463461, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.528246] env[61728]: DEBUG nova.network.neutron [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Updating instance_info_cache with network_info: [{"id": "590df666-2609-450c-b0e2-e5df320ae98b", "address": "fa:16:3e:a5:d2:e9", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap590df666-26", "ovs_interfaceid": "590df666-2609-450c-b0e2-e5df320ae98b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 521.610978] env[61728]: DEBUG nova.compute.manager [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 521.673413] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463462, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065337} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.673749] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 521.674715] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7d1bfe-cb65-40bb-98f7-9b07948508f9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.701753] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Reconfiguring VM instance instance-00000007 to attach disk [datastore1] a3e9af38-5115-4337-a20e-5d6c6ec655fe/a3e9af38-5115-4337-a20e-5d6c6ec655fe.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 521.702652] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e05f571-a80f-4713-91eb-7f70efe13550 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.722938] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 521.722938] env[61728]: value = "task-463463" [ 521.722938] env[61728]: _type = "Task" [ 521.722938] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.731457] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463463, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.744398] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance a0831461-ece1-43ee-92f6-34d7d4e673e2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 521.982599] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463461, 'name': CreateVM_Task, 'duration_secs': 0.77975} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.982792] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 521.983500] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.983706] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.984042] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 521.984306] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09fbdc9a-b676-485e-a041-24d2d95b172d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.989546] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 521.989546] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d5a85a-4843-dc2f-e399-86fbd2e3abf2" [ 521.989546] env[61728]: _type = "Task" [ 521.989546] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.999062] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d5a85a-4843-dc2f-e399-86fbd2e3abf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.034952] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "refresh_cache-4247e356-ef10-41b3-bb77-106103134d59" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 522.035343] env[61728]: DEBUG nova.compute.manager [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Instance network_info: |[{"id": "590df666-2609-450c-b0e2-e5df320ae98b", "address": "fa:16:3e:a5:d2:e9", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap590df666-26", "ovs_interfaceid": "590df666-2609-450c-b0e2-e5df320ae98b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 522.035774] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:d2:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd733acc2-07d0-479e-918c-ec8a21925389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '590df666-2609-450c-b0e2-e5df320ae98b', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 522.044931] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Creating folder: Project (08367aaceba548fe93faaedf6371817d). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 522.045088] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ddd9608-bb59-4712-b61a-3a8f35e7f886 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.055291] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Created folder: Project (08367aaceba548fe93faaedf6371817d) in parent group-v121913. [ 522.055291] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Creating folder: Instances. Parent ref: group-v121942. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 522.055475] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-69fcb841-833e-4d3f-b25d-fdda22cb6a12 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.063932] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Created folder: Instances in parent group-v121942. [ 522.064428] env[61728]: DEBUG oslo.service.loopingcall [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 522.064644] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 522.064915] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-333b3dca-9735-4584-a94a-462886831d6d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.087252] env[61728]: DEBUG nova.compute.manager [req-64f17ced-5442-4672-8810-70cb534fbf61 req-933e2e36-e135-4f68-b205-208549ba180a service nova] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Received event network-vif-plugged-590df666-2609-450c-b0e2-e5df320ae98b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 522.087475] env[61728]: DEBUG oslo_concurrency.lockutils [req-64f17ced-5442-4672-8810-70cb534fbf61 req-933e2e36-e135-4f68-b205-208549ba180a service nova] Acquiring lock "4247e356-ef10-41b3-bb77-106103134d59-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.087754] env[61728]: DEBUG oslo_concurrency.lockutils [req-64f17ced-5442-4672-8810-70cb534fbf61 req-933e2e36-e135-4f68-b205-208549ba180a service nova] Lock "4247e356-ef10-41b3-bb77-106103134d59-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.087942] env[61728]: DEBUG oslo_concurrency.lockutils [req-64f17ced-5442-4672-8810-70cb534fbf61 req-933e2e36-e135-4f68-b205-208549ba180a service nova] Lock "4247e356-ef10-41b3-bb77-106103134d59-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.088122] env[61728]: DEBUG nova.compute.manager [req-64f17ced-5442-4672-8810-70cb534fbf61 req-933e2e36-e135-4f68-b205-208549ba180a service nova] [instance: 4247e356-ef10-41b3-bb77-106103134d59] No waiting events found dispatching network-vif-plugged-590df666-2609-450c-b0e2-e5df320ae98b {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 522.088356] env[61728]: WARNING nova.compute.manager [req-64f17ced-5442-4672-8810-70cb534fbf61 req-933e2e36-e135-4f68-b205-208549ba180a service nova] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Received unexpected event network-vif-plugged-590df666-2609-450c-b0e2-e5df320ae98b for instance with vm_state building and task_state spawning. [ 522.093500] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 522.093500] env[61728]: value = "task-463466" [ 522.093500] env[61728]: _type = "Task" [ 522.093500] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.102769] env[61728]: DEBUG nova.network.neutron [-] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.103854] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463466, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.150378] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.234082] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463463, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.247850] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance be014f77-53ca-42a5-9255-2ba950cca314 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 522.325132] env[61728]: DEBUG nova.compute.manager [None req-18cee19b-8fa2-4596-87e9-a3703facb807 tempest-ServerExternalEventsTest-1489248873 tempest-ServerExternalEventsTest-1489248873-project] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Received event network-changed {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 522.325328] env[61728]: DEBUG nova.compute.manager [None req-18cee19b-8fa2-4596-87e9-a3703facb807 tempest-ServerExternalEventsTest-1489248873 tempest-ServerExternalEventsTest-1489248873-project] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Refreshing instance network info cache due to event network-changed. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 522.325547] env[61728]: DEBUG oslo_concurrency.lockutils [None req-18cee19b-8fa2-4596-87e9-a3703facb807 tempest-ServerExternalEventsTest-1489248873 tempest-ServerExternalEventsTest-1489248873-project] Acquiring lock "refresh_cache-08ad3f4a-12a8-48d3-8829-f6744648a95e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.325689] env[61728]: DEBUG oslo_concurrency.lockutils [None req-18cee19b-8fa2-4596-87e9-a3703facb807 tempest-ServerExternalEventsTest-1489248873 tempest-ServerExternalEventsTest-1489248873-project] Acquired lock "refresh_cache-08ad3f4a-12a8-48d3-8829-f6744648a95e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.325841] env[61728]: DEBUG nova.network.neutron [None req-18cee19b-8fa2-4596-87e9-a3703facb807 tempest-ServerExternalEventsTest-1489248873 tempest-ServerExternalEventsTest-1489248873-project] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 522.443591] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Acquiring lock "3cd416c6-9926-45ee-afe3-d3468e4e06c2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.443941] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Lock "3cd416c6-9926-45ee-afe3-d3468e4e06c2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.444247] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Acquiring lock "3cd416c6-9926-45ee-afe3-d3468e4e06c2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.444440] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Lock "3cd416c6-9926-45ee-afe3-d3468e4e06c2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.444618] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Lock "3cd416c6-9926-45ee-afe3-d3468e4e06c2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.447182] env[61728]: INFO nova.compute.manager [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Terminating instance [ 522.450052] env[61728]: DEBUG nova.compute.manager [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 522.450270] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 522.451099] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38755e53-ea51-4f44-80a8-be218f2b5d88 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.467793] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 522.468197] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc40b980-0540-4bd9-9ada-d0f268ae7d54 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.474896] env[61728]: DEBUG oslo_vmware.api [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Waiting for the task: (returnval){ [ 522.474896] env[61728]: value = "task-463468" [ 522.474896] env[61728]: _type = "Task" [ 522.474896] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.484329] env[61728]: DEBUG oslo_vmware.api [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463468, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.500618] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d5a85a-4843-dc2f-e399-86fbd2e3abf2, 'name': SearchDatastore_Task, 'duration_secs': 0.014815} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.500998] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 522.501234] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 522.501489] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.501654] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.501867] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 522.502109] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf1530b6-831d-4a6b-b1c4-c8362d114062 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.511991] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 522.512222] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 522.513040] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc754b01-838f-48ca-a87e-99302ab01901 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.519214] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 522.519214] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527f6850-7bb9-31c1-5817-3fc6bd52c391" [ 522.519214] env[61728]: _type = "Task" [ 522.519214] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.527573] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527f6850-7bb9-31c1-5817-3fc6bd52c391, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.610415] env[61728]: INFO nova.compute.manager [-] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Took 1.26 seconds to deallocate network for instance. [ 522.610641] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463466, 'name': CreateVM_Task, 'duration_secs': 0.465608} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.613246] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 522.614130] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.614343] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.614665] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 522.619480] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81bc4989-c974-4e12-acc6-3adad680088c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.625722] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 522.625722] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5248f077-1d50-689c-91f0-f26579645ca3" [ 522.625722] env[61728]: _type = "Task" [ 522.625722] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.636614] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5248f077-1d50-689c-91f0-f26579645ca3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.738253] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463463, 'name': ReconfigVM_Task, 'duration_secs': 0.624524} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.738253] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Reconfigured VM instance instance-00000007 to attach disk [datastore1] a3e9af38-5115-4337-a20e-5d6c6ec655fe/a3e9af38-5115-4337-a20e-5d6c6ec655fe.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 522.738253] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d8ea4d7c-a772-4478-af69-4827716e3493 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.747643] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 522.747643] env[61728]: value = "task-463469" [ 522.747643] env[61728]: _type = "Task" [ 522.747643] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.753334] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance db8e373d-e06c-43f0-a79c-2c1aaf65a01a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 522.763287] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463469, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.986657] env[61728]: DEBUG oslo_vmware.api [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463468, 'name': PowerOffVM_Task, 'duration_secs': 0.236567} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.986926] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 522.988099] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 522.988099] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-93440e20-3af8-416b-aaf3-6f67b7fd430d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.033279] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527f6850-7bb9-31c1-5817-3fc6bd52c391, 'name': SearchDatastore_Task, 'duration_secs': 0.011479} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.033279] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9e89ced-16a1-45c6-9293-effbfaa76120 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.040703] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 523.040703] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52626ebf-13d6-b795-6f71-e64fac906c48" [ 523.040703] env[61728]: _type = "Task" [ 523.040703] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.048586] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52626ebf-13d6-b795-6f71-e64fac906c48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.061636] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 523.061636] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 523.061636] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Deleting the datastore file [datastore1] 3cd416c6-9926-45ee-afe3-d3468e4e06c2 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 523.061636] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ffbdd4ce-2cbd-44ff-a9ac-c47e55b4b026 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.068447] env[61728]: DEBUG oslo_vmware.api [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Waiting for the task: (returnval){ [ 523.068447] env[61728]: value = "task-463471" [ 523.068447] env[61728]: _type = "Task" [ 523.068447] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.076541] env[61728]: DEBUG oslo_vmware.api [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463471, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.123657] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.139502] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5248f077-1d50-689c-91f0-f26579645ca3, 'name': SearchDatastore_Task, 'duration_secs': 0.010554} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.139502] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.139502] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 523.139502] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.258773] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance c8f1f4ae-ab08-4a03-a3fd-94014509e7e7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 523.258773] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 523.258972] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 523.261286] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463469, 'name': Rename_Task, 'duration_secs': 0.247266} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.262557] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 523.262557] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a603e008-2dbc-462b-a84a-f3bba01c4138 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.269158] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 523.269158] env[61728]: value = "task-463472" [ 523.269158] env[61728]: _type = "Task" [ 523.269158] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.278964] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463472, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.368139] env[61728]: DEBUG nova.network.neutron [None req-18cee19b-8fa2-4596-87e9-a3703facb807 tempest-ServerExternalEventsTest-1489248873 tempest-ServerExternalEventsTest-1489248873-project] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Updating instance_info_cache with network_info: [{"id": "e083e0fd-486c-40fb-b54d-c58808ccf0a3", "address": "fa:16:3e:05:49:1a", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape083e0fd-48", "ovs_interfaceid": "e083e0fd-486c-40fb-b54d-c58808ccf0a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.552342] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52626ebf-13d6-b795-6f71-e64fac906c48, 'name': SearchDatastore_Task, 'duration_secs': 0.013725} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.557260] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.557549] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618/ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 523.558177] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.558585] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 523.558804] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dec9be97-c1fb-43e2-bec5-84d39143598b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.560891] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9efde115-cd9c-4d78-b292-4ff0f9c2dfaa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.572819] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 523.572819] env[61728]: value = "task-463473" [ 523.572819] env[61728]: _type = "Task" [ 523.572819] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.574514] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 523.574514] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 523.579420] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea4cb917-09ea-4c25-a347-5df99b476136 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.592322] env[61728]: DEBUG oslo_vmware.api [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Task: {'id': task-463471, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.400016} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.593342] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 523.593540] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 523.593739] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 523.593942] env[61728]: INFO nova.compute.manager [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Took 1.14 seconds to destroy the instance on the hypervisor. [ 523.594709] env[61728]: DEBUG oslo.service.loopingcall [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 523.594998] env[61728]: DEBUG nova.compute.manager [-] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 523.595132] env[61728]: DEBUG nova.network.neutron [-] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 523.604406] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463473, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.604845] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 523.604845] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520c111f-5f10-b6fc-acc5-721934e1c682" [ 523.604845] env[61728]: _type = "Task" [ 523.604845] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.618557] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520c111f-5f10-b6fc-acc5-721934e1c682, 'name': SearchDatastore_Task, 'duration_secs': 0.013524} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.619514] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76421092-aeae-4560-a8e4-de0d1cb1ac0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.628179] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 523.628179] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ca6faf-62e1-e9c3-3676-29b6777a4e2d" [ 523.628179] env[61728]: _type = "Task" [ 523.628179] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.636813] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ca6faf-62e1-e9c3-3676-29b6777a4e2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.666321] env[61728]: DEBUG nova.compute.manager [req-308743d3-00cb-4ef2-adab-381584765c28 req-54b82c21-836f-41b1-bc4e-a3295f7c9d58 service nova] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Received event network-changed-86d2fa33-d261-42a1-b49d-1fd478f8311b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 523.666582] env[61728]: DEBUG nova.compute.manager [req-308743d3-00cb-4ef2-adab-381584765c28 req-54b82c21-836f-41b1-bc4e-a3295f7c9d58 service nova] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Refreshing instance network info cache due to event network-changed-86d2fa33-d261-42a1-b49d-1fd478f8311b. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 523.666813] env[61728]: DEBUG oslo_concurrency.lockutils [req-308743d3-00cb-4ef2-adab-381584765c28 req-54b82c21-836f-41b1-bc4e-a3295f7c9d58 service nova] Acquiring lock "refresh_cache-ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.667018] env[61728]: DEBUG oslo_concurrency.lockutils [req-308743d3-00cb-4ef2-adab-381584765c28 req-54b82c21-836f-41b1-bc4e-a3295f7c9d58 service nova] Acquired lock "refresh_cache-ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.667319] env[61728]: DEBUG nova.network.neutron [req-308743d3-00cb-4ef2-adab-381584765c28 req-54b82c21-836f-41b1-bc4e-a3295f7c9d58 service nova] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Refreshing network info cache for port 86d2fa33-d261-42a1-b49d-1fd478f8311b {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 523.685277] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c25949f-a024-461f-b9d8-dc29cd7dd896 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.694106] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed865c6-91f8-484b-9a3a-5129422cbebf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.731660] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f244ba-3686-41f7-bdd0-bdde8dc05fab {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.741746] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97bee7aa-6ef3-4678-a2a4-cee45850a0a8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.754802] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 523.781249] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463472, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.872843] env[61728]: DEBUG oslo_concurrency.lockutils [None req-18cee19b-8fa2-4596-87e9-a3703facb807 tempest-ServerExternalEventsTest-1489248873 tempest-ServerExternalEventsTest-1489248873-project] Releasing lock "refresh_cache-08ad3f4a-12a8-48d3-8829-f6744648a95e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.085604] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463473, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.143163] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ca6faf-62e1-e9c3-3676-29b6777a4e2d, 'name': SearchDatastore_Task, 'duration_secs': 0.031885} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.143900] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.144177] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 4247e356-ef10-41b3-bb77-106103134d59/4247e356-ef10-41b3-bb77-106103134d59.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 524.144449] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2c44170-2067-4a51-a464-f22d593d08e4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.154912] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 524.154912] env[61728]: value = "task-463475" [ 524.154912] env[61728]: _type = "Task" [ 524.154912] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.167814] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463475, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.258114] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 524.291962] env[61728]: DEBUG oslo_vmware.api [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463472, 'name': PowerOnVM_Task, 'duration_secs': 0.729593} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.292297] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 524.292514] env[61728]: INFO nova.compute.manager [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Took 11.35 seconds to spawn the instance on the hypervisor. [ 524.292705] env[61728]: DEBUG nova.compute.manager [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 524.293565] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5519af9f-05b4-4c59-aa20-60e6af20fdae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.516514] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Acquiring lock "08ad3f4a-12a8-48d3-8829-f6744648a95e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.516779] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Lock "08ad3f4a-12a8-48d3-8829-f6744648a95e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.517535] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Acquiring lock "08ad3f4a-12a8-48d3-8829-f6744648a95e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.517657] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Lock "08ad3f4a-12a8-48d3-8829-f6744648a95e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.517762] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Lock "08ad3f4a-12a8-48d3-8829-f6744648a95e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.520228] env[61728]: INFO nova.compute.manager [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Terminating instance [ 524.522419] env[61728]: DEBUG nova.compute.manager [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 524.522670] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 524.523507] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95e358c-f8f2-485b-a5f4-7a575d4d44f3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.533438] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 524.533438] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75f7902b-62c9-4a17-9e9f-bc86c62df9ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.541704] env[61728]: DEBUG oslo_vmware.api [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Waiting for the task: (returnval){ [ 524.541704] env[61728]: value = "task-463476" [ 524.541704] env[61728]: _type = "Task" [ 524.541704] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.552157] env[61728]: DEBUG oslo_vmware.api [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463476, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.587505] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463473, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.908531} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.588049] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618/ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 524.588225] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 524.588583] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6c0672e-46f1-4804-bbd9-d4134c86b7d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.599108] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 524.599108] env[61728]: value = "task-463477" [ 524.599108] env[61728]: _type = "Task" [ 524.599108] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.610572] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463477, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.631028] env[61728]: DEBUG nova.network.neutron [req-308743d3-00cb-4ef2-adab-381584765c28 req-54b82c21-836f-41b1-bc4e-a3295f7c9d58 service nova] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Updated VIF entry in instance network info cache for port 86d2fa33-d261-42a1-b49d-1fd478f8311b. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 524.631771] env[61728]: DEBUG nova.network.neutron [req-308743d3-00cb-4ef2-adab-381584765c28 req-54b82c21-836f-41b1-bc4e-a3295f7c9d58 service nova] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Updating instance_info_cache with network_info: [{"id": "86d2fa33-d261-42a1-b49d-1fd478f8311b", "address": "fa:16:3e:f5:2d:48", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86d2fa33-d2", "ovs_interfaceid": "86d2fa33-d261-42a1-b49d-1fd478f8311b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.666520] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463475, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.764758] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 524.766416] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.620s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.766416] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.240s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.767013] env[61728]: INFO nova.compute.claims [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 524.815490] env[61728]: DEBUG nova.network.neutron [-] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.822423] env[61728]: INFO nova.compute.manager [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Took 20.10 seconds to build instance. [ 525.062051] env[61728]: DEBUG oslo_vmware.api [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463476, 'name': PowerOffVM_Task, 'duration_secs': 0.260785} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.062051] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 525.062051] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 525.062051] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-446a8714-abd8-4d7c-8dc9-3e8e8fc50b65 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.112129] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463477, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084686} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.112451] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 525.113533] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cefebb33-6069-4590-b8fc-ed2116591e21 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.140558] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Reconfiguring VM instance instance-00000008 to attach disk [datastore1] ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618/ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 525.141872] env[61728]: DEBUG oslo_concurrency.lockutils [req-308743d3-00cb-4ef2-adab-381584765c28 req-54b82c21-836f-41b1-bc4e-a3295f7c9d58 service nova] Releasing lock "refresh_cache-ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.141872] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a092eee2-b196-4dd8-8109-4e8e0844fa5a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.172852] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463475, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.175998] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 525.175998] env[61728]: value = "task-463479" [ 525.175998] env[61728]: _type = "Task" [ 525.175998] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.186378] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463479, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.321922] env[61728]: INFO nova.compute.manager [-] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Took 1.73 seconds to deallocate network for instance. [ 525.328953] env[61728]: DEBUG oslo_concurrency.lockutils [None req-070c9746-4dbd-4285-b0a1-679dd58add09 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "a3e9af38-5115-4337-a20e-5d6c6ec655fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.616s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.443577] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 525.443886] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 525.444098] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Deleting the datastore file [datastore1] 08ad3f4a-12a8-48d3-8829-f6744648a95e {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 525.445113] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1f0d5a2-017e-46cf-a3df-dbfedd6d7b03 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.453142] env[61728]: DEBUG oslo_vmware.api [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Waiting for the task: (returnval){ [ 525.453142] env[61728]: value = "task-463480" [ 525.453142] env[61728]: _type = "Task" [ 525.453142] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.462031] env[61728]: DEBUG oslo_vmware.api [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463480, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.496635] env[61728]: DEBUG nova.compute.manager [req-ebee0f1d-87d7-442c-a7ea-f0f04ecb6a06 req-e1bd984f-94b2-4432-9dd6-f605758969ea service nova] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Received event network-changed-590df666-2609-450c-b0e2-e5df320ae98b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 525.496829] env[61728]: DEBUG nova.compute.manager [req-ebee0f1d-87d7-442c-a7ea-f0f04ecb6a06 req-e1bd984f-94b2-4432-9dd6-f605758969ea service nova] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Refreshing instance network info cache due to event network-changed-590df666-2609-450c-b0e2-e5df320ae98b. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 525.497052] env[61728]: DEBUG oslo_concurrency.lockutils [req-ebee0f1d-87d7-442c-a7ea-f0f04ecb6a06 req-e1bd984f-94b2-4432-9dd6-f605758969ea service nova] Acquiring lock "refresh_cache-4247e356-ef10-41b3-bb77-106103134d59" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.497196] env[61728]: DEBUG oslo_concurrency.lockutils [req-ebee0f1d-87d7-442c-a7ea-f0f04ecb6a06 req-e1bd984f-94b2-4432-9dd6-f605758969ea service nova] Acquired lock "refresh_cache-4247e356-ef10-41b3-bb77-106103134d59" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.497355] env[61728]: DEBUG nova.network.neutron [req-ebee0f1d-87d7-442c-a7ea-f0f04ecb6a06 req-e1bd984f-94b2-4432-9dd6-f605758969ea service nova] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Refreshing network info cache for port 590df666-2609-450c-b0e2-e5df320ae98b {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 525.676296] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463475, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.233408} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.680382] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 4247e356-ef10-41b3-bb77-106103134d59/4247e356-ef10-41b3-bb77-106103134d59.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 525.680592] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 525.680866] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-10704430-4591-47a9-8110-ae3076e96927 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.689830] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463479, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.691884] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 525.691884] env[61728]: value = "task-463482" [ 525.691884] env[61728]: _type = "Task" [ 525.691884] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.702047] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463482, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.837825] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.838338] env[61728]: DEBUG nova.compute.manager [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 525.969177] env[61728]: DEBUG oslo_vmware.api [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Task: {'id': task-463480, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.402642} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.970029] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 525.970029] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 525.970143] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 525.970246] env[61728]: INFO nova.compute.manager [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Took 1.45 seconds to destroy the instance on the hypervisor. [ 525.970482] env[61728]: DEBUG oslo.service.loopingcall [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 525.970670] env[61728]: DEBUG nova.compute.manager [-] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 525.970760] env[61728]: DEBUG nova.network.neutron [-] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 526.190839] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463479, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.202107] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463482, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.152338} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.202315] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 526.203229] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df22cdd5-39bd-4c68-ac18-655102413899 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.230246] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Reconfiguring VM instance instance-00000009 to attach disk [datastore1] 4247e356-ef10-41b3-bb77-106103134d59/4247e356-ef10-41b3-bb77-106103134d59.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 526.235930] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-426dafdc-ff3e-41c8-8907-2c7645c19150 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.260891] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 526.260891] env[61728]: value = "task-463483" [ 526.260891] env[61728]: _type = "Task" [ 526.260891] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.276366] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463483, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.284023] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d5a872e-bea5-4d82-b36b-50d0bf57db14 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.291919] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8bf1e7-0264-4673-a9b3-a718251f2263 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.324392] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cdace9e-a26e-47e8-bfe5-6a8a0724a664 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.334769] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ccff44-5eb9-4fbb-bcbf-146ce71697b6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.355738] env[61728]: DEBUG nova.compute.provider_tree [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 526.372606] env[61728]: DEBUG oslo_concurrency.lockutils [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.493848] env[61728]: DEBUG nova.network.neutron [req-ebee0f1d-87d7-442c-a7ea-f0f04ecb6a06 req-e1bd984f-94b2-4432-9dd6-f605758969ea service nova] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Updated VIF entry in instance network info cache for port 590df666-2609-450c-b0e2-e5df320ae98b. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 526.494222] env[61728]: DEBUG nova.network.neutron [req-ebee0f1d-87d7-442c-a7ea-f0f04ecb6a06 req-e1bd984f-94b2-4432-9dd6-f605758969ea service nova] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Updating instance_info_cache with network_info: [{"id": "590df666-2609-450c-b0e2-e5df320ae98b", "address": "fa:16:3e:a5:d2:e9", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap590df666-26", "ovs_interfaceid": "590df666-2609-450c-b0e2-e5df320ae98b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.692693] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463479, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.773162] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463483, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.895254] env[61728]: ERROR nova.scheduler.client.report [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [req-f52f2f3f-97d1-4c3e-bd0a-326e83ab9a23] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f52f2f3f-97d1-4c3e-bd0a-326e83ab9a23"}]} [ 526.919494] env[61728]: DEBUG nova.scheduler.client.report [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 526.941437] env[61728]: DEBUG nova.scheduler.client.report [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 526.941684] env[61728]: DEBUG nova.compute.provider_tree [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 526.957419] env[61728]: DEBUG nova.scheduler.client.report [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 526.980456] env[61728]: DEBUG nova.scheduler.client.report [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 526.997126] env[61728]: DEBUG oslo_concurrency.lockutils [req-ebee0f1d-87d7-442c-a7ea-f0f04ecb6a06 req-e1bd984f-94b2-4432-9dd6-f605758969ea service nova] Releasing lock "refresh_cache-4247e356-ef10-41b3-bb77-106103134d59" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 526.998244] env[61728]: DEBUG nova.compute.manager [req-ebee0f1d-87d7-442c-a7ea-f0f04ecb6a06 req-e1bd984f-94b2-4432-9dd6-f605758969ea service nova] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Received event network-vif-deleted-95c872e2-49de-44f2-bc09-b8595260a29f {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 527.168787] env[61728]: DEBUG nova.network.neutron [-] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.197019] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463479, 'name': ReconfigVM_Task, 'duration_secs': 2.004178} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.199982] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Reconfigured VM instance instance-00000008 to attach disk [datastore1] ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618/ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 527.201020] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2347b25b-6442-464c-bdef-9a5b386cc135 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.214985] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 527.214985] env[61728]: value = "task-463485" [ 527.214985] env[61728]: _type = "Task" [ 527.214985] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.227176] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463485, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.279008] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463483, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.376468] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701ae7bd-fd66-484a-b8ec-0c21f4fa03b7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.384543] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc4c832e-43f4-494b-8743-e72198b673b9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.424400] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4f16b2-a972-4b6e-91c3-394682810745 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.432326] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c99512-3d0f-4575-919d-982268716058 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.448072] env[61728]: DEBUG nova.compute.provider_tree [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 527.672412] env[61728]: INFO nova.compute.manager [-] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Took 1.70 seconds to deallocate network for instance. [ 527.729686] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463485, 'name': Rename_Task, 'duration_secs': 0.169246} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.729961] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 527.731440] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2c94d7e4-6a79-4c85-8ded-48ee0a485d13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.740172] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 527.740172] env[61728]: value = "task-463487" [ 527.740172] env[61728]: _type = "Task" [ 527.740172] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.752995] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463487, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.779542] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463483, 'name': ReconfigVM_Task, 'duration_secs': 1.303521} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.779865] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Reconfigured VM instance instance-00000009 to attach disk [datastore1] 4247e356-ef10-41b3-bb77-106103134d59/4247e356-ef10-41b3-bb77-106103134d59.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 527.780583] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fe25554b-5687-4ea8-8279-b2672232eb8c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.791166] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 527.791166] env[61728]: value = "task-463488" [ 527.791166] env[61728]: _type = "Task" [ 527.791166] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.801752] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463488, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.973067] env[61728]: ERROR nova.scheduler.client.report [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [req-4559d8f3-8949-402f-b9b2-39d6770ca5d8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4559d8f3-8949-402f-b9b2-39d6770ca5d8"}]} [ 527.996331] env[61728]: DEBUG nova.scheduler.client.report [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 528.025959] env[61728]: DEBUG nova.scheduler.client.report [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 528.026429] env[61728]: DEBUG nova.compute.provider_tree [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 528.050076] env[61728]: DEBUG nova.scheduler.client.report [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 528.122134] env[61728]: DEBUG nova.scheduler.client.report [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 528.184246] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.256581] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463487, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.270380] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Acquiring lock "aec920d1-12fb-4d41-9663-b4c2b7e142d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.270380] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Lock "aec920d1-12fb-4d41-9663-b4c2b7e142d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.307285] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463488, 'name': Rename_Task, 'duration_secs': 0.192666} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.307285] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 528.307285] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-781c58d7-f6d0-4a78-9610-4ddb316753f8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.315256] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 528.315256] env[61728]: value = "task-463489" [ 528.315256] env[61728]: _type = "Task" [ 528.315256] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.331597] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463489, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.393402] env[61728]: DEBUG nova.compute.manager [req-d75580a9-0b61-4933-8646-1fdd0a24a639 req-f0912aa5-9a78-47d8-9fce-96314bda68b3 service nova] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Received event network-vif-deleted-e083e0fd-486c-40fb-b54d-c58808ccf0a3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 528.530127] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4829e240-66cc-48cd-afa4-de573ad21f27 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.538419] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3eb99fc-098f-4620-8c8e-53754301ca1c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.572137] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356cba3b-afa4-4de6-9776-b6108ac36935 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.580755] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c97b86ee-d793-4e53-98cb-fbf11cfd9450 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.597670] env[61728]: DEBUG nova.compute.provider_tree [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.754527] env[61728]: DEBUG oslo_vmware.api [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463487, 'name': PowerOnVM_Task, 'duration_secs': 0.661525} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.754834] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 528.754980] env[61728]: INFO nova.compute.manager [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Took 12.20 seconds to spawn the instance on the hypervisor. [ 528.755427] env[61728]: DEBUG nova.compute.manager [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 528.756493] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f89939f7-6980-493a-8a25-a7c2fa9ea552 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.831949] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463489, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.105329] env[61728]: DEBUG nova.scheduler.client.report [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 529.227126] env[61728]: DEBUG nova.compute.manager [req-9f3a4347-c1a9-4ec1-8fb0-89ddf30b8af2 req-2efab221-364b-48b5-918f-1a2629219335 service nova] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Received event network-vif-deleted-fd11f13f-a69b-43da-bc74-f5d74871fe0e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 529.286969] env[61728]: INFO nova.compute.manager [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Took 23.84 seconds to build instance. [ 529.328630] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463489, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.614632] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.847s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.614632] env[61728]: DEBUG nova.compute.manager [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 529.618430] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.307s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.620381] env[61728]: INFO nova.compute.claims [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.791682] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0334d0ff-23ed-49c8-b4e6-d2ac75b6b2f8 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.366s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.827580] env[61728]: DEBUG oslo_vmware.api [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463489, 'name': PowerOnVM_Task, 'duration_secs': 1.270036} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.827902] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 529.828115] env[61728]: INFO nova.compute.manager [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Took 10.64 seconds to spawn the instance on the hypervisor. [ 529.828480] env[61728]: DEBUG nova.compute.manager [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 529.829647] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d6acab-6698-428e-8df2-fc37f42f46b4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.128637] env[61728]: DEBUG nova.compute.utils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 530.134807] env[61728]: DEBUG nova.compute.manager [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 530.134807] env[61728]: DEBUG nova.network.neutron [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 530.298021] env[61728]: DEBUG nova.compute.manager [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 530.325162] env[61728]: DEBUG nova.policy [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2000460ed1e24feba059663cb49b4a21', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f532dea1a024d94bd2bc700ffc35936', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 530.351523] env[61728]: INFO nova.compute.manager [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Took 24.40 seconds to build instance. [ 530.635569] env[61728]: DEBUG nova.compute.manager [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 530.766948] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "2d389d5e-a53f-424e-90af-5ed05e2f93c7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.767230] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "2d389d5e-a53f-424e-90af-5ed05e2f93c7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.823778] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.857346] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7635480b-f0b0-453e-8448-92f3ee926e65 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "4247e356-ef10-41b3-bb77-106103134d59" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.915s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.090460] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d13e74-c69f-4815-a1c4-541e4e03c62c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.107202] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-844ca6c2-386b-4054-b534-fde1d0ea9546 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.152289] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fae5e30-4edb-4dbc-ab7c-552b8229e1c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.163302] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411668b7-66ce-4dba-89a4-1eef041fb459 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.181439] env[61728]: DEBUG nova.compute.provider_tree [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 531.251520] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Acquiring lock "4e5dc620-68a1-4e10-a8be-702c4999ca10" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.251733] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Lock "4e5dc620-68a1-4e10-a8be-702c4999ca10" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.293342] env[61728]: DEBUG nova.network.neutron [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Successfully created port: 94e809f2-e1ed-45f6-872f-ec7c798028aa {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 531.360690] env[61728]: DEBUG nova.compute.manager [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 531.654234] env[61728]: DEBUG nova.compute.manager [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 531.686024] env[61728]: DEBUG nova.scheduler.client.report [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.693413] env[61728]: DEBUG nova.virt.hardware [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 531.693413] env[61728]: DEBUG nova.virt.hardware [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 531.693413] env[61728]: DEBUG nova.virt.hardware [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.693623] env[61728]: DEBUG nova.virt.hardware [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 531.693623] env[61728]: DEBUG nova.virt.hardware [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.693623] env[61728]: DEBUG nova.virt.hardware [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 531.694930] env[61728]: DEBUG nova.virt.hardware [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 531.694930] env[61728]: DEBUG nova.virt.hardware [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 531.694930] env[61728]: DEBUG nova.virt.hardware [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 531.694930] env[61728]: DEBUG nova.virt.hardware [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 531.694930] env[61728]: DEBUG nova.virt.hardware [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 531.696908] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2478315c-9c8e-4bd6-beab-b61669cc9769 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.708395] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6ba3dc-aa6a-48ae-992f-0b7898df1da4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.900054] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.201822] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.584s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.203242] env[61728]: DEBUG nova.compute.manager [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 532.205415] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.158s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.206835] env[61728]: INFO nova.compute.claims [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 532.712188] env[61728]: DEBUG nova.compute.utils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.721111] env[61728]: DEBUG nova.compute.manager [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 532.721111] env[61728]: DEBUG nova.network.neutron [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 532.769912] env[61728]: DEBUG nova.policy [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0777be3ba7d5466ea41ba7ee98e1f535', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f587ea8b8fe432aafd3e0ecba054bd0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 533.219596] env[61728]: DEBUG nova.compute.manager [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 533.343398] env[61728]: DEBUG nova.network.neutron [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Successfully created port: da4f38d4-527c-47c4-9fa7-a5009a0a4e6c {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 533.514205] env[61728]: DEBUG nova.network.neutron [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Successfully updated port: 94e809f2-e1ed-45f6-872f-ec7c798028aa {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 533.523796] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "4247e356-ef10-41b3-bb77-106103134d59" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.523796] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "4247e356-ef10-41b3-bb77-106103134d59" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.523796] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "4247e356-ef10-41b3-bb77-106103134d59-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.523796] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "4247e356-ef10-41b3-bb77-106103134d59-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.524292] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "4247e356-ef10-41b3-bb77-106103134d59-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.526648] env[61728]: INFO nova.compute.manager [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Terminating instance [ 533.529982] env[61728]: DEBUG nova.compute.manager [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 533.531010] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 533.535136] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a283967f-10f6-4c34-acf0-453bab9fd4c9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.547245] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 533.547585] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dcd80ca5-7f16-486f-a9a5-8e6260de72af {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.558844] env[61728]: DEBUG oslo_vmware.api [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 533.558844] env[61728]: value = "task-463494" [ 533.558844] env[61728]: _type = "Task" [ 533.558844] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.582086] env[61728]: DEBUG oslo_vmware.api [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463494, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.657018] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.657309] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.721575] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf492581-7610-4ad7-b6b8-b91bf9a2fbbf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.735348] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d1040ec-f64a-447a-a318-185957b60780 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.776179] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12eca0c-0ec5-4b97-b966-c79143ea6e1e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.786511] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ee4255-cec4-4d61-90a7-5781ac514a84 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.804253] env[61728]: DEBUG nova.compute.provider_tree [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 534.018208] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Acquiring lock "refresh_cache-b86bf217-d1f8-4d37-aa87-3a2d27d70420" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.018362] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Acquired lock "refresh_cache-b86bf217-d1f8-4d37-aa87-3a2d27d70420" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.018521] env[61728]: DEBUG nova.network.neutron [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 534.076558] env[61728]: DEBUG oslo_vmware.api [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463494, 'name': PowerOffVM_Task, 'duration_secs': 0.216711} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.076558] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 534.076558] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 534.076778] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-255ce6ca-e709-4482-a45c-37ede58f9d8d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.178533] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 534.180041] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 534.180041] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleting the datastore file [datastore1] 4247e356-ef10-41b3-bb77-106103134d59 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 534.180212] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b1ab45d-ea2e-4177-8f1e-f3e8d4958bed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.191898] env[61728]: DEBUG oslo_vmware.api [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 534.191898] env[61728]: value = "task-463497" [ 534.191898] env[61728]: _type = "Task" [ 534.191898] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.202856] env[61728]: DEBUG oslo_vmware.api [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463497, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.236799] env[61728]: DEBUG nova.compute.manager [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 534.270428] env[61728]: DEBUG nova.virt.hardware [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 534.273637] env[61728]: DEBUG nova.virt.hardware [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 534.273906] env[61728]: DEBUG nova.virt.hardware [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 534.274118] env[61728]: DEBUG nova.virt.hardware [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 534.274282] env[61728]: DEBUG nova.virt.hardware [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 534.274451] env[61728]: DEBUG nova.virt.hardware [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 534.274685] env[61728]: DEBUG nova.virt.hardware [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 534.275073] env[61728]: DEBUG nova.virt.hardware [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 534.275073] env[61728]: DEBUG nova.virt.hardware [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 534.275178] env[61728]: DEBUG nova.virt.hardware [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 534.275605] env[61728]: DEBUG nova.virt.hardware [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 534.276336] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4834a0f9-2349-49c5-b38f-70550231245d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.286837] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88551de5-23dc-45e5-8220-1e54fab26ce4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.307654] env[61728]: DEBUG nova.scheduler.client.report [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 534.588345] env[61728]: DEBUG nova.network.neutron [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.707220] env[61728]: DEBUG oslo_vmware.api [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463497, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.466514} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.707282] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 534.707422] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 534.707597] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 534.707763] env[61728]: INFO nova.compute.manager [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Took 1.18 seconds to destroy the instance on the hypervisor. [ 534.708025] env[61728]: DEBUG oslo.service.loopingcall [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 534.708214] env[61728]: DEBUG nova.compute.manager [-] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 534.708434] env[61728]: DEBUG nova.network.neutron [-] [instance: 4247e356-ef10-41b3-bb77-106103134d59] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 534.740786] env[61728]: DEBUG nova.compute.manager [req-54b69165-eba3-4a65-9a10-05a17c9426df req-e4b0167f-f5bf-4ff8-99dc-ab391e18817f service nova] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Received event network-vif-plugged-94e809f2-e1ed-45f6-872f-ec7c798028aa {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 534.741248] env[61728]: DEBUG oslo_concurrency.lockutils [req-54b69165-eba3-4a65-9a10-05a17c9426df req-e4b0167f-f5bf-4ff8-99dc-ab391e18817f service nova] Acquiring lock "b86bf217-d1f8-4d37-aa87-3a2d27d70420-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.741248] env[61728]: DEBUG oslo_concurrency.lockutils [req-54b69165-eba3-4a65-9a10-05a17c9426df req-e4b0167f-f5bf-4ff8-99dc-ab391e18817f service nova] Lock "b86bf217-d1f8-4d37-aa87-3a2d27d70420-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.741521] env[61728]: DEBUG oslo_concurrency.lockutils [req-54b69165-eba3-4a65-9a10-05a17c9426df req-e4b0167f-f5bf-4ff8-99dc-ab391e18817f service nova] Lock "b86bf217-d1f8-4d37-aa87-3a2d27d70420-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.741579] env[61728]: DEBUG nova.compute.manager [req-54b69165-eba3-4a65-9a10-05a17c9426df req-e4b0167f-f5bf-4ff8-99dc-ab391e18817f service nova] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] No waiting events found dispatching network-vif-plugged-94e809f2-e1ed-45f6-872f-ec7c798028aa {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 534.743862] env[61728]: WARNING nova.compute.manager [req-54b69165-eba3-4a65-9a10-05a17c9426df req-e4b0167f-f5bf-4ff8-99dc-ab391e18817f service nova] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Received unexpected event network-vif-plugged-94e809f2-e1ed-45f6-872f-ec7c798028aa for instance with vm_state building and task_state spawning. [ 534.767504] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "fc3c1d93-b095-4d80-8d71-243b66d85a46" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.767996] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "fc3c1d93-b095-4d80-8d71-243b66d85a46" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.815161] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.815161] env[61728]: DEBUG nova.compute.manager [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 534.817248] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.748s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.818868] env[61728]: INFO nova.compute.claims [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 535.060695] env[61728]: DEBUG nova.network.neutron [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Updating instance_info_cache with network_info: [{"id": "94e809f2-e1ed-45f6-872f-ec7c798028aa", "address": "fa:16:3e:8b:05:d2", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94e809f2-e1", "ovs_interfaceid": "94e809f2-e1ed-45f6-872f-ec7c798028aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.323194] env[61728]: DEBUG nova.compute.utils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 535.327138] env[61728]: DEBUG nova.compute.manager [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 535.327319] env[61728]: DEBUG nova.network.neutron [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 535.541690] env[61728]: DEBUG nova.policy [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0777be3ba7d5466ea41ba7ee98e1f535', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f587ea8b8fe432aafd3e0ecba054bd0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 535.564114] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Releasing lock "refresh_cache-b86bf217-d1f8-4d37-aa87-3a2d27d70420" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.564479] env[61728]: DEBUG nova.compute.manager [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Instance network_info: |[{"id": "94e809f2-e1ed-45f6-872f-ec7c798028aa", "address": "fa:16:3e:8b:05:d2", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94e809f2-e1", "ovs_interfaceid": "94e809f2-e1ed-45f6-872f-ec7c798028aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 535.564892] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:05:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '457c42cd-4ddb-4374-923e-d419b7f6eaff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94e809f2-e1ed-45f6-872f-ec7c798028aa', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 535.576600] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Creating folder: Project (2f532dea1a024d94bd2bc700ffc35936). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 535.576600] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e381a799-e2c1-403b-9117-656df5eae82f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.589473] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Created folder: Project (2f532dea1a024d94bd2bc700ffc35936) in parent group-v121913. [ 535.589712] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Creating folder: Instances. Parent ref: group-v121947. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 535.589923] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2c51840b-1951-4858-b7dd-11578a841cf2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.600207] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Created folder: Instances in parent group-v121947. [ 535.601258] env[61728]: DEBUG oslo.service.loopingcall [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.601258] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 535.601258] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-463dbd78-ed5f-452f-8965-6040633469aa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.616969] env[61728]: DEBUG nova.network.neutron [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Successfully updated port: da4f38d4-527c-47c4-9fa7-a5009a0a4e6c {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 535.628032] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 535.628032] env[61728]: value = "task-463500" [ 535.628032] env[61728]: _type = "Task" [ 535.628032] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.636133] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463500, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.828272] env[61728]: DEBUG nova.compute.manager [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 535.943920] env[61728]: DEBUG nova.network.neutron [-] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.983185] env[61728]: DEBUG nova.network.neutron [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Successfully created port: 7af87a49-cf84-45ac-8bb4-e6d6062e4db0 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 536.124055] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "refresh_cache-adb58292-62cd-4142-ad66-db014bcbdc49" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.124397] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquired lock "refresh_cache-adb58292-62cd-4142-ad66-db014bcbdc49" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.124544] env[61728]: DEBUG nova.network.neutron [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 536.145020] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463500, 'name': CreateVM_Task, 'duration_secs': 0.408155} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.145020] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 536.146575] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.146740] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.147066] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 536.147590] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74bd342f-bacb-44c6-a6e0-17a5b67fc9ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.154397] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Waiting for the task: (returnval){ [ 536.154397] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52312e77-0547-4e48-6b93-55121459693b" [ 536.154397] env[61728]: _type = "Task" [ 536.154397] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.173830] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52312e77-0547-4e48-6b93-55121459693b, 'name': SearchDatastore_Task, 'duration_secs': 0.009546} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.173954] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.174763] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 536.174763] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.174763] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.174763] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 536.175085] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d2a3c98-8c02-4f1e-b767-bc8fc547cd4a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.188060] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 536.188060] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 536.188183] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fcb204a-c618-4a6c-9761-ae8910535001 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.194818] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Waiting for the task: (returnval){ [ 536.194818] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522e5150-b8ac-177b-b717-d7570fc7fc59" [ 536.194818] env[61728]: _type = "Task" [ 536.194818] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.211025] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522e5150-b8ac-177b-b717-d7570fc7fc59, 'name': SearchDatastore_Task, 'duration_secs': 0.008552} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.211948] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc29e5eb-ac89-462a-afc4-bacd60b9425a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.221934] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Waiting for the task: (returnval){ [ 536.221934] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b21a1e-445b-fc72-be25-e3488e90a53c" [ 536.221934] env[61728]: _type = "Task" [ 536.221934] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.231198] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b21a1e-445b-fc72-be25-e3488e90a53c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.324799] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7cd2c42-bc62-4b09-95dc-dbabeb859348 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.332764] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e7de73e-4213-4f70-b486-8e98664b1049 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.372282] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513b9875-7be5-46ff-a9d2-3e8f6cfe9fa5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.385473] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d263f3ce-9b6b-4a8e-9867-b89a458aaa60 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.403582] env[61728]: DEBUG nova.compute.provider_tree [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 536.448051] env[61728]: INFO nova.compute.manager [-] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Took 1.74 seconds to deallocate network for instance. [ 536.527786] env[61728]: DEBUG oslo_concurrency.lockutils [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Acquiring lock "37b8ca5a-e0f8-414b-a363-e56db520f027" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.527786] env[61728]: DEBUG oslo_concurrency.lockutils [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Lock "37b8ca5a-e0f8-414b-a363-e56db520f027" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.686454] env[61728]: DEBUG nova.network.neutron [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.737113] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b21a1e-445b-fc72-be25-e3488e90a53c, 'name': SearchDatastore_Task, 'duration_secs': 0.008477} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.739615] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.740304] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b86bf217-d1f8-4d37-aa87-3a2d27d70420/b86bf217-d1f8-4d37-aa87-3a2d27d70420.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 536.740304] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c4e150a-5e2e-4a47-bcd7-39d2e4b02394 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.749656] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Waiting for the task: (returnval){ [ 536.749656] env[61728]: value = "task-463502" [ 536.749656] env[61728]: _type = "Task" [ 536.749656] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.762854] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463502, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.841559] env[61728]: DEBUG nova.compute.manager [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 536.858263] env[61728]: DEBUG nova.network.neutron [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Updating instance_info_cache with network_info: [{"id": "da4f38d4-527c-47c4-9fa7-a5009a0a4e6c", "address": "fa:16:3e:2e:c1:36", "network": {"id": "4da787ba-7f1e-46fb-9902-d90ec8854a65", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-249168995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f587ea8b8fe432aafd3e0ecba054bd0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48937bd1-23dc-413f-b46b-59bf9e709aa0", "external-id": "nsx-vlan-transportzone-160", "segmentation_id": 160, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda4f38d4-52", "ovs_interfaceid": "da4f38d4-527c-47c4-9fa7-a5009a0a4e6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.871665] env[61728]: DEBUG nova.virt.hardware [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 536.871992] env[61728]: DEBUG nova.virt.hardware [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 536.872184] env[61728]: DEBUG nova.virt.hardware [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.872626] env[61728]: DEBUG nova.virt.hardware [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 536.872626] env[61728]: DEBUG nova.virt.hardware [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.872760] env[61728]: DEBUG nova.virt.hardware [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 536.872875] env[61728]: DEBUG nova.virt.hardware [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 536.873030] env[61728]: DEBUG nova.virt.hardware [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 536.873453] env[61728]: DEBUG nova.virt.hardware [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 536.873453] env[61728]: DEBUG nova.virt.hardware [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 536.873571] env[61728]: DEBUG nova.virt.hardware [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 536.874423] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3877caaa-69ca-41a5-9f07-a8e9bcf924ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.884352] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a36cc92-2bd4-4ecb-87c8-6664865a948c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.907758] env[61728]: DEBUG nova.scheduler.client.report [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 536.957658] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.183063] env[61728]: INFO nova.compute.manager [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Rebuilding instance [ 537.239824] env[61728]: DEBUG nova.compute.manager [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 537.241138] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e6d456-3869-4588-ba63-9452ca7cb9b0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.263729] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463502, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490587} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.263729] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b86bf217-d1f8-4d37-aa87-3a2d27d70420/b86bf217-d1f8-4d37-aa87-3a2d27d70420.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 537.263729] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 537.263729] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-394c20b9-205d-4372-89b4-674f4b742726 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.272452] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Waiting for the task: (returnval){ [ 537.272452] env[61728]: value = "task-463503" [ 537.272452] env[61728]: _type = "Task" [ 537.272452] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.283389] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463503, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.301187] env[61728]: DEBUG nova.compute.manager [req-a2ca95f0-df2a-457d-850a-4d9ab290a5a0 req-825c3f3e-e3a3-4b4f-977f-38b9aafea40c service nova] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Received event network-vif-plugged-da4f38d4-527c-47c4-9fa7-a5009a0a4e6c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 537.301797] env[61728]: DEBUG oslo_concurrency.lockutils [req-a2ca95f0-df2a-457d-850a-4d9ab290a5a0 req-825c3f3e-e3a3-4b4f-977f-38b9aafea40c service nova] Acquiring lock "adb58292-62cd-4142-ad66-db014bcbdc49-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.301797] env[61728]: DEBUG oslo_concurrency.lockutils [req-a2ca95f0-df2a-457d-850a-4d9ab290a5a0 req-825c3f3e-e3a3-4b4f-977f-38b9aafea40c service nova] Lock "adb58292-62cd-4142-ad66-db014bcbdc49-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.301976] env[61728]: DEBUG oslo_concurrency.lockutils [req-a2ca95f0-df2a-457d-850a-4d9ab290a5a0 req-825c3f3e-e3a3-4b4f-977f-38b9aafea40c service nova] Lock "adb58292-62cd-4142-ad66-db014bcbdc49-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.302042] env[61728]: DEBUG nova.compute.manager [req-a2ca95f0-df2a-457d-850a-4d9ab290a5a0 req-825c3f3e-e3a3-4b4f-977f-38b9aafea40c service nova] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] No waiting events found dispatching network-vif-plugged-da4f38d4-527c-47c4-9fa7-a5009a0a4e6c {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 537.302189] env[61728]: WARNING nova.compute.manager [req-a2ca95f0-df2a-457d-850a-4d9ab290a5a0 req-825c3f3e-e3a3-4b4f-977f-38b9aafea40c service nova] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Received unexpected event network-vif-plugged-da4f38d4-527c-47c4-9fa7-a5009a0a4e6c for instance with vm_state building and task_state spawning. [ 537.363797] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Releasing lock "refresh_cache-adb58292-62cd-4142-ad66-db014bcbdc49" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.363797] env[61728]: DEBUG nova.compute.manager [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Instance network_info: |[{"id": "da4f38d4-527c-47c4-9fa7-a5009a0a4e6c", "address": "fa:16:3e:2e:c1:36", "network": {"id": "4da787ba-7f1e-46fb-9902-d90ec8854a65", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-249168995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f587ea8b8fe432aafd3e0ecba054bd0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48937bd1-23dc-413f-b46b-59bf9e709aa0", "external-id": "nsx-vlan-transportzone-160", "segmentation_id": 160, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda4f38d4-52", "ovs_interfaceid": "da4f38d4-527c-47c4-9fa7-a5009a0a4e6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 537.364256] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:c1:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48937bd1-23dc-413f-b46b-59bf9e709aa0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'da4f38d4-527c-47c4-9fa7-a5009a0a4e6c', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 537.374721] env[61728]: DEBUG oslo.service.loopingcall [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 537.375511] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 537.378274] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bc4a4426-5bc5-4664-bfbf-82db76240feb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.398397] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 537.398397] env[61728]: value = "task-463504" [ 537.398397] env[61728]: _type = "Task" [ 537.398397] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.410965] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463504, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.412949] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.594s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.412949] env[61728]: DEBUG nova.compute.manager [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 537.416045] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.279s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.417678] env[61728]: INFO nova.compute.claims [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 537.764445] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 537.764822] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d4c2d7b6-5168-4967-b0e7-5f5a777ac6aa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.779092] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 537.779092] env[61728]: value = "task-463505" [ 537.779092] env[61728]: _type = "Task" [ 537.779092] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.787044] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463503, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070298} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.787993] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 537.789370] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec59ff8-4db8-44d6-8eec-29c6d8c7afc1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.799573] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463505, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.826314] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] b86bf217-d1f8-4d37-aa87-3a2d27d70420/b86bf217-d1f8-4d37-aa87-3a2d27d70420.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 537.827024] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c74534f2-c712-445a-91ed-ec620e8c5a9d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.851119] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Waiting for the task: (returnval){ [ 537.851119] env[61728]: value = "task-463506" [ 537.851119] env[61728]: _type = "Task" [ 537.851119] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.861753] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463506, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.915084] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463504, 'name': CreateVM_Task, 'duration_secs': 0.418099} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.918269] env[61728]: DEBUG nova.compute.utils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 537.918977] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 537.919642] env[61728]: DEBUG nova.compute.manager [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 537.919926] env[61728]: DEBUG nova.network.neutron [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 537.925033] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.925033] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.925033] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 537.925676] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cef7349-57af-491c-bc02-963922ab39bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.933463] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Acquiring lock "f947d808-ebea-4593-b8e6-d3851c5216dc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.933463] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lock "f947d808-ebea-4593-b8e6-d3851c5216dc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.936967] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 537.936967] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5270efb1-e690-8522-5982-b4718dcaf374" [ 537.936967] env[61728]: _type = "Task" [ 537.936967] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.949673] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5270efb1-e690-8522-5982-b4718dcaf374, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.952831] env[61728]: DEBUG nova.network.neutron [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Successfully updated port: 7af87a49-cf84-45ac-8bb4-e6d6062e4db0 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 538.004267] env[61728]: DEBUG nova.policy [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aac08b5321344d5ab384a3714e6e0bc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '383f38592de14b4ba93fa9a6eeec40aa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 538.293984] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463505, 'name': PowerOffVM_Task, 'duration_secs': 0.266035} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.294761] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 538.294761] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 538.295876] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f9b8c9-aaf7-4ad1-a5d6-ed48295b7037 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.303426] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 538.303653] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9733d57e-48cd-4367-8378-fde7de4dabdf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.312306] env[61728]: DEBUG nova.network.neutron [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Successfully created port: 132601da-9151-425d-afd3-3878b0424d7e {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 538.361902] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463506, 'name': ReconfigVM_Task, 'duration_secs': 0.303057} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.363649] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Reconfigured VM instance instance-0000000a to attach disk [datastore1] b86bf217-d1f8-4d37-aa87-3a2d27d70420/b86bf217-d1f8-4d37-aa87-3a2d27d70420.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 538.363649] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-163ec353-9466-4cdf-82b5-d98b157e893e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.372692] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Waiting for the task: (returnval){ [ 538.372692] env[61728]: value = "task-463508" [ 538.372692] env[61728]: _type = "Task" [ 538.372692] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.386174] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463508, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.389294] env[61728]: DEBUG nova.compute.manager [req-f4d3956b-cc04-4328-accc-e939b13aa8ca req-738a2be8-db9b-477c-b373-db5dc6a46f04 service nova] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Received event network-changed-94e809f2-e1ed-45f6-872f-ec7c798028aa {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 538.389378] env[61728]: DEBUG nova.compute.manager [req-f4d3956b-cc04-4328-accc-e939b13aa8ca req-738a2be8-db9b-477c-b373-db5dc6a46f04 service nova] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Refreshing instance network info cache due to event network-changed-94e809f2-e1ed-45f6-872f-ec7c798028aa. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 538.389628] env[61728]: DEBUG oslo_concurrency.lockutils [req-f4d3956b-cc04-4328-accc-e939b13aa8ca req-738a2be8-db9b-477c-b373-db5dc6a46f04 service nova] Acquiring lock "refresh_cache-b86bf217-d1f8-4d37-aa87-3a2d27d70420" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.389765] env[61728]: DEBUG oslo_concurrency.lockutils [req-f4d3956b-cc04-4328-accc-e939b13aa8ca req-738a2be8-db9b-477c-b373-db5dc6a46f04 service nova] Acquired lock "refresh_cache-b86bf217-d1f8-4d37-aa87-3a2d27d70420" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.389923] env[61728]: DEBUG nova.network.neutron [req-f4d3956b-cc04-4328-accc-e939b13aa8ca req-738a2be8-db9b-477c-b373-db5dc6a46f04 service nova] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Refreshing network info cache for port 94e809f2-e1ed-45f6-872f-ec7c798028aa {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 538.392161] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 538.392388] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 538.392581] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleting the datastore file [datastore1] ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 538.393643] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63e03193-2e2c-43ad-aa60-6b47a4cfb553 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.408772] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 538.408772] env[61728]: value = "task-463509" [ 538.408772] env[61728]: _type = "Task" [ 538.408772] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.421850] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463509, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.421850] env[61728]: DEBUG nova.compute.manager [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 538.453273] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5270efb1-e690-8522-5982-b4718dcaf374, 'name': SearchDatastore_Task, 'duration_secs': 0.010851} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.453601] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.453849] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 538.454690] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.454690] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.454690] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 538.454880] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ef2dd2e8-1d23-4d70-9958-43b27d1fa848 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.458312] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "refresh_cache-e56a59a6-fc0f-45ca-810a-76b18bae173a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.458559] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquired lock "refresh_cache-e56a59a6-fc0f-45ca-810a-76b18bae173a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.458559] env[61728]: DEBUG nova.network.neutron [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 538.480589] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 538.480589] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 538.480589] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25e7cf27-a655-4281-b4e3-043441953c83 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.488296] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 538.488296] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522b5326-9f98-ee65-f72d-a1f5b9b853e4" [ 538.488296] env[61728]: _type = "Task" [ 538.488296] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.506354] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522b5326-9f98-ee65-f72d-a1f5b9b853e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.885110] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463508, 'name': Rename_Task, 'duration_secs': 0.182594} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.885392] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 538.888670] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b79a5d1e-f97b-45d2-a7cb-a9deb382ac52 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.898014] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Waiting for the task: (returnval){ [ 538.898014] env[61728]: value = "task-463510" [ 538.898014] env[61728]: _type = "Task" [ 538.898014] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.909832] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463510, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.929324] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463509, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169975} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.929324] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 538.929478] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 538.929572] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 538.940512] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-997659ca-e0b1-40fd-82fc-2c3d44d7134a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.953069] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c827268-ccfd-4c85-9a6c-7e996ed88bd8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.988557] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c3ab2c2-65b0-4ba4-adca-47cd4f168c8a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.001657] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522b5326-9f98-ee65-f72d-a1f5b9b853e4, 'name': SearchDatastore_Task, 'duration_secs': 0.013336} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.005100] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6711b38-ab4e-426b-9766-c532501072d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.008518] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eda8645-8070-4b39-8920-f3558ab347ea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.023692] env[61728]: DEBUG nova.compute.provider_tree [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 539.026052] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 539.026052] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52951f5e-73fe-1cad-afd4-78d58227f4b4" [ 539.026052] env[61728]: _type = "Task" [ 539.026052] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.035821] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52951f5e-73fe-1cad-afd4-78d58227f4b4, 'name': SearchDatastore_Task, 'duration_secs': 0.009748} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.036590] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.036803] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] adb58292-62cd-4142-ad66-db014bcbdc49/adb58292-62cd-4142-ad66-db014bcbdc49.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 539.039213] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-31d2b608-9c45-4f90-90ab-f149e30b362f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.040320] env[61728]: DEBUG nova.network.neutron [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.048480] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 539.048480] env[61728]: value = "task-463511" [ 539.048480] env[61728]: _type = "Task" [ 539.048480] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.058160] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463511, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.249631] env[61728]: DEBUG nova.network.neutron [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Updating instance_info_cache with network_info: [{"id": "7af87a49-cf84-45ac-8bb4-e6d6062e4db0", "address": "fa:16:3e:04:17:42", "network": {"id": "4da787ba-7f1e-46fb-9902-d90ec8854a65", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-249168995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f587ea8b8fe432aafd3e0ecba054bd0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48937bd1-23dc-413f-b46b-59bf9e709aa0", "external-id": "nsx-vlan-transportzone-160", "segmentation_id": 160, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7af87a49-cf", "ovs_interfaceid": "7af87a49-cf84-45ac-8bb4-e6d6062e4db0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.410286] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463510, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.432265] env[61728]: DEBUG nova.compute.manager [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 539.481907] env[61728]: DEBUG nova.virt.hardware [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 539.482349] env[61728]: DEBUG nova.virt.hardware [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 539.482542] env[61728]: DEBUG nova.virt.hardware [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 539.482753] env[61728]: DEBUG nova.virt.hardware [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 539.482906] env[61728]: DEBUG nova.virt.hardware [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 539.483068] env[61728]: DEBUG nova.virt.hardware [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 539.483598] env[61728]: DEBUG nova.virt.hardware [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 539.483793] env[61728]: DEBUG nova.virt.hardware [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 539.483974] env[61728]: DEBUG nova.virt.hardware [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 539.484161] env[61728]: DEBUG nova.virt.hardware [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 539.484658] env[61728]: DEBUG nova.virt.hardware [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 539.485364] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0d7e39-ead0-4b72-8243-53728d0bb7bf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.498028] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2e177a-0cc8-44aa-9047-771da410e7dc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.520182] env[61728]: DEBUG nova.network.neutron [req-f4d3956b-cc04-4328-accc-e939b13aa8ca req-738a2be8-db9b-477c-b373-db5dc6a46f04 service nova] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Updated VIF entry in instance network info cache for port 94e809f2-e1ed-45f6-872f-ec7c798028aa. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 539.520525] env[61728]: DEBUG nova.network.neutron [req-f4d3956b-cc04-4328-accc-e939b13aa8ca req-738a2be8-db9b-477c-b373-db5dc6a46f04 service nova] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Updating instance_info_cache with network_info: [{"id": "94e809f2-e1ed-45f6-872f-ec7c798028aa", "address": "fa:16:3e:8b:05:d2", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94e809f2-e1", "ovs_interfaceid": "94e809f2-e1ed-45f6-872f-ec7c798028aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.527899] env[61728]: DEBUG nova.scheduler.client.report [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 539.562743] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463511, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481676} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.562743] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] adb58292-62cd-4142-ad66-db014bcbdc49/adb58292-62cd-4142-ad66-db014bcbdc49.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 539.562743] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 539.562743] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09eceea8-2cd2-4f44-b7e5-5249300cc97a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.569192] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 539.569192] env[61728]: value = "task-463512" [ 539.569192] env[61728]: _type = "Task" [ 539.569192] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.578483] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463512, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.752651] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Releasing lock "refresh_cache-e56a59a6-fc0f-45ca-810a-76b18bae173a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.752651] env[61728]: DEBUG nova.compute.manager [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Instance network_info: |[{"id": "7af87a49-cf84-45ac-8bb4-e6d6062e4db0", "address": "fa:16:3e:04:17:42", "network": {"id": "4da787ba-7f1e-46fb-9902-d90ec8854a65", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-249168995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f587ea8b8fe432aafd3e0ecba054bd0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48937bd1-23dc-413f-b46b-59bf9e709aa0", "external-id": "nsx-vlan-transportzone-160", "segmentation_id": 160, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7af87a49-cf", "ovs_interfaceid": "7af87a49-cf84-45ac-8bb4-e6d6062e4db0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 539.753071] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:17:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48937bd1-23dc-413f-b46b-59bf9e709aa0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7af87a49-cf84-45ac-8bb4-e6d6062e4db0', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 539.761676] env[61728]: DEBUG oslo.service.loopingcall [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 539.762032] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 539.762151] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1fa3e34a-c7d8-4882-9446-fe53e28d7fd4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.785200] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 539.785200] env[61728]: value = "task-463513" [ 539.785200] env[61728]: _type = "Task" [ 539.785200] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.794742] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463513, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.916034] env[61728]: DEBUG oslo_vmware.api [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463510, 'name': PowerOnVM_Task, 'duration_secs': 0.520432} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.918168] env[61728]: DEBUG nova.network.neutron [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Successfully updated port: 132601da-9151-425d-afd3-3878b0424d7e {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 539.918214] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 539.918421] env[61728]: INFO nova.compute.manager [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Took 8.26 seconds to spawn the instance on the hypervisor. [ 539.921584] env[61728]: DEBUG nova.compute.manager [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 539.922550] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48488877-80c2-4247-aef7-94d3340f8c1a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.981949] env[61728]: DEBUG nova.virt.hardware [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 539.982923] env[61728]: DEBUG nova.virt.hardware [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 539.983215] env[61728]: DEBUG nova.virt.hardware [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 539.983471] env[61728]: DEBUG nova.virt.hardware [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 539.983669] env[61728]: DEBUG nova.virt.hardware [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 539.984487] env[61728]: DEBUG nova.virt.hardware [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 539.984487] env[61728]: DEBUG nova.virt.hardware [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 539.984487] env[61728]: DEBUG nova.virt.hardware [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 539.984487] env[61728]: DEBUG nova.virt.hardware [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 539.984764] env[61728]: DEBUG nova.virt.hardware [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 539.984764] env[61728]: DEBUG nova.virt.hardware [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 539.987097] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ebd3a5b-a856-465d-aab2-6a59d0ef6709 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.998756] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b0f34a-5d05-4f1d-accf-2f4187312a6f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.014204] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f5:2d:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c492f5cc-7ae0-4cab-823c-0d5dd8c60b26', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '86d2fa33-d261-42a1-b49d-1fd478f8311b', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 540.023152] env[61728]: DEBUG oslo.service.loopingcall [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.023152] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 540.024114] env[61728]: DEBUG oslo_concurrency.lockutils [req-f4d3956b-cc04-4328-accc-e939b13aa8ca req-738a2be8-db9b-477c-b373-db5dc6a46f04 service nova] Releasing lock "refresh_cache-b86bf217-d1f8-4d37-aa87-3a2d27d70420" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.024114] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-84cd76e6-3c5f-4fea-8d91-f005eb8d3e76 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.038712] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.623s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.039367] env[61728]: DEBUG nova.compute.manager [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 540.042199] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 19.161s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.050892] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 540.050892] env[61728]: value = "task-463514" [ 540.050892] env[61728]: _type = "Task" [ 540.050892] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.063330] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463514, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.086188] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463512, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081415} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.086188] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 540.086188] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d138da-941b-4399-924b-636599fc04b0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.113226] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Reconfiguring VM instance instance-0000000b to attach disk [datastore1] adb58292-62cd-4142-ad66-db014bcbdc49/adb58292-62cd-4142-ad66-db014bcbdc49.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 540.113894] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8ed2035-3cf2-4de4-9e2a-13c3bbef00ed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.135021] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 540.135021] env[61728]: value = "task-463515" [ 540.135021] env[61728]: _type = "Task" [ 540.135021] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.143978] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463515, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.296735] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463513, 'name': CreateVM_Task, 'duration_secs': 0.383197} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.296846] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 540.297598] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.297764] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.298098] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 540.298356] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19230bf7-ea9c-41c7-8b64-738b4157db3f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.305656] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 540.305656] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b03048-c837-3ba4-dc1e-af9b0e9d0c0d" [ 540.305656] env[61728]: _type = "Task" [ 540.305656] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.317412] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b03048-c837-3ba4-dc1e-af9b0e9d0c0d, 'name': SearchDatastore_Task, 'duration_secs': 0.010682} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.318588] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.318588] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 540.318588] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.318588] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.318808] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 540.318808] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2dfcb94-7385-4639-acfc-8299a7501026 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.330200] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 540.330200] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 540.330200] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a3d6ac9-8619-4f70-ba27-f143ed1879ec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.335547] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 540.335547] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522b24da-2b70-3a38-bb09-557e8f586142" [ 540.335547] env[61728]: _type = "Task" [ 540.335547] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.343744] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522b24da-2b70-3a38-bb09-557e8f586142, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.423122] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "refresh_cache-3c4f7d53-6ba6-4e2c-b324-e651640b304b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.423374] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquired lock "refresh_cache-3c4f7d53-6ba6-4e2c-b324-e651640b304b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.423423] env[61728]: DEBUG nova.network.neutron [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 540.443302] env[61728]: INFO nova.compute.manager [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Took 33.93 seconds to build instance. [ 540.548559] env[61728]: INFO nova.compute.claims [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 540.558058] env[61728]: DEBUG nova.compute.utils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 540.558058] env[61728]: DEBUG nova.compute.manager [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 540.558058] env[61728]: DEBUG nova.network.neutron [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 540.566176] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463514, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.598327] env[61728]: DEBUG nova.policy [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e51efcf23db4d5496edbc3b7104b4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32bb35e1dfce40e48be08bb568d3f2b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 540.646603] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463515, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.848066] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522b24da-2b70-3a38-bb09-557e8f586142, 'name': SearchDatastore_Task, 'duration_secs': 0.009356} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.849221] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ca187cc-dfaa-4bcd-81c9-edcd0384c813 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.857027] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 540.857027] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d2769a-8c93-aebc-9ed3-86567f792ee9" [ 540.857027] env[61728]: _type = "Task" [ 540.857027] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.870261] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d2769a-8c93-aebc-9ed3-86567f792ee9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.944363] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50fb97ad-9145-4c8f-8412-34e0c89b9037 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Lock "b86bf217-d1f8-4d37-aa87-3a2d27d70420" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.446s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.952821] env[61728]: DEBUG nova.network.neutron [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Successfully created port: edc90742-f02d-42ad-85cd-b4242f797c33 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 541.011416] env[61728]: DEBUG nova.network.neutron [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.058170] env[61728]: INFO nova.compute.resource_tracker [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating resource usage from migration 04adf437-6f01-45eb-848e-3da8ff14cd18 [ 541.062641] env[61728]: DEBUG nova.compute.manager [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 541.085253] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463514, 'name': CreateVM_Task, 'duration_secs': 0.915992} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.085917] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 541.087302] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.087764] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.088427] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 541.088782] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b97cd318-379b-470d-9265-bb5d72429854 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.095560] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 541.095560] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5235599a-4b73-20d6-1851-3832f085028c" [ 541.095560] env[61728]: _type = "Task" [ 541.095560] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.106443] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5235599a-4b73-20d6-1851-3832f085028c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.146829] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463515, 'name': ReconfigVM_Task, 'duration_secs': 0.956528} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.150045] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Reconfigured VM instance instance-0000000b to attach disk [datastore1] adb58292-62cd-4142-ad66-db014bcbdc49/adb58292-62cd-4142-ad66-db014bcbdc49.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 541.151186] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bdee0235-a971-4f2f-8a01-7aff03b72cd1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.159407] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 541.159407] env[61728]: value = "task-463516" [ 541.159407] env[61728]: _type = "Task" [ 541.159407] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.173871] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463516, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.369250] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d2769a-8c93-aebc-9ed3-86567f792ee9, 'name': SearchDatastore_Task, 'duration_secs': 0.01147} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.372117] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.372877] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] e56a59a6-fc0f-45ca-810a-76b18bae173a/e56a59a6-fc0f-45ca-810a-76b18bae173a.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 541.372877] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-088eb9f6-dedf-44be-b739-bfc82cf239a5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.381078] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 541.381078] env[61728]: value = "task-463517" [ 541.381078] env[61728]: _type = "Task" [ 541.381078] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.393020] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463517, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.451674] env[61728]: DEBUG nova.compute.manager [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 541.458280] env[61728]: DEBUG nova.network.neutron [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Updating instance_info_cache with network_info: [{"id": "132601da-9151-425d-afd3-3878b0424d7e", "address": "fa:16:3e:94:a8:a2", "network": {"id": "17a8ea01-7320-4089-9138-a6fd17ac4054", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-599506352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "383f38592de14b4ba93fa9a6eeec40aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap132601da-91", "ovs_interfaceid": "132601da-9151-425d-afd3-3878b0424d7e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.504153] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd9626f-b368-42ec-961a-fda5964c3bc9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.516025] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca18dc8e-4e39-4595-827f-173687419308 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.560432] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da054ab7-f22e-40a3-b349-15c3a05778a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.568895] env[61728]: DEBUG nova.compute.manager [req-96322bc4-93ad-4bfa-b168-adf9edeeedcb req-5ec928bb-0b57-48af-90a5-f105cc245be7 service nova] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Received event network-vif-deleted-590df666-2609-450c-b0e2-e5df320ae98b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 541.568895] env[61728]: DEBUG nova.compute.manager [req-96322bc4-93ad-4bfa-b168-adf9edeeedcb req-5ec928bb-0b57-48af-90a5-f105cc245be7 service nova] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Received event network-changed-da4f38d4-527c-47c4-9fa7-a5009a0a4e6c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 541.568895] env[61728]: DEBUG nova.compute.manager [req-96322bc4-93ad-4bfa-b168-adf9edeeedcb req-5ec928bb-0b57-48af-90a5-f105cc245be7 service nova] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Refreshing instance network info cache due to event network-changed-da4f38d4-527c-47c4-9fa7-a5009a0a4e6c. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 541.568895] env[61728]: DEBUG oslo_concurrency.lockutils [req-96322bc4-93ad-4bfa-b168-adf9edeeedcb req-5ec928bb-0b57-48af-90a5-f105cc245be7 service nova] Acquiring lock "refresh_cache-adb58292-62cd-4142-ad66-db014bcbdc49" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.568895] env[61728]: DEBUG oslo_concurrency.lockutils [req-96322bc4-93ad-4bfa-b168-adf9edeeedcb req-5ec928bb-0b57-48af-90a5-f105cc245be7 service nova] Acquired lock "refresh_cache-adb58292-62cd-4142-ad66-db014bcbdc49" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.569710] env[61728]: DEBUG nova.network.neutron [req-96322bc4-93ad-4bfa-b168-adf9edeeedcb req-5ec928bb-0b57-48af-90a5-f105cc245be7 service nova] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Refreshing network info cache for port da4f38d4-527c-47c4-9fa7-a5009a0a4e6c {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 541.585447] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84b567c-251d-4323-acbe-f4c81c21f0cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.604136] env[61728]: DEBUG nova.compute.provider_tree [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 541.619026] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5235599a-4b73-20d6-1851-3832f085028c, 'name': SearchDatastore_Task, 'duration_secs': 0.010093} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.620592] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.620934] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 541.621220] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.621376] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.621620] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 541.622211] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6220b915-2fbd-41d3-b59a-f41c424b3c87 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.640734] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 541.640899] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 541.642583] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c97a108-68a3-4c25-a21a-f84bd2d3cdfd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.651715] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 541.651715] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a88423-44ef-c1e2-53e9-b59779e53a67" [ 541.651715] env[61728]: _type = "Task" [ 541.651715] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.663962] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a88423-44ef-c1e2-53e9-b59779e53a67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.675298] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463516, 'name': Rename_Task, 'duration_secs': 0.315848} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.675720] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 541.676054] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36ea336e-ecb2-4971-98f9-82e632ae8593 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.685425] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 541.685425] env[61728]: value = "task-463518" [ 541.685425] env[61728]: _type = "Task" [ 541.685425] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.697179] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463518, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.906473] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463517, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478337} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.906749] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] e56a59a6-fc0f-45ca-810a-76b18bae173a/e56a59a6-fc0f-45ca-810a-76b18bae173a.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 541.908027] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 541.908027] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-47785856-7ad0-4ee2-b02c-a7b635a5a286 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.921180] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 541.921180] env[61728]: value = "task-463519" [ 541.921180] env[61728]: _type = "Task" [ 541.921180] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.940943] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463519, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.967488] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Releasing lock "refresh_cache-3c4f7d53-6ba6-4e2c-b324-e651640b304b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.967860] env[61728]: DEBUG nova.compute.manager [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Instance network_info: |[{"id": "132601da-9151-425d-afd3-3878b0424d7e", "address": "fa:16:3e:94:a8:a2", "network": {"id": "17a8ea01-7320-4089-9138-a6fd17ac4054", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-599506352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "383f38592de14b4ba93fa9a6eeec40aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap132601da-91", "ovs_interfaceid": "132601da-9151-425d-afd3-3878b0424d7e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 541.971056] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:a8:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '132601da-9151-425d-afd3-3878b0424d7e', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 541.979610] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Creating folder: Project (383f38592de14b4ba93fa9a6eeec40aa). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 541.979916] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-23f2c72f-812b-471c-8ea9-e2250ba0427e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.990905] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.994701] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Created folder: Project (383f38592de14b4ba93fa9a6eeec40aa) in parent group-v121913. [ 541.994861] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Creating folder: Instances. Parent ref: group-v121953. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 541.995251] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ebf341d9-d3fa-4dff-adf8-6d13da6fc7c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.010419] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Created folder: Instances in parent group-v121953. [ 542.010799] env[61728]: DEBUG oslo.service.loopingcall [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.011101] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 542.011400] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-42d1ab2b-6da7-4650-b144-2f593c1159be {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.044949] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 542.044949] env[61728]: value = "task-463522" [ 542.044949] env[61728]: _type = "Task" [ 542.044949] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.053902] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463522, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.084873] env[61728]: DEBUG nova.compute.manager [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 542.107571] env[61728]: DEBUG nova.scheduler.client.report [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 542.114680] env[61728]: DEBUG nova.virt.hardware [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 542.114680] env[61728]: DEBUG nova.virt.hardware [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 542.114816] env[61728]: DEBUG nova.virt.hardware [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 542.115396] env[61728]: DEBUG nova.virt.hardware [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 542.115396] env[61728]: DEBUG nova.virt.hardware [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 542.115396] env[61728]: DEBUG nova.virt.hardware [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 542.115662] env[61728]: DEBUG nova.virt.hardware [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 542.115662] env[61728]: DEBUG nova.virt.hardware [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 542.115828] env[61728]: DEBUG nova.virt.hardware [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 542.115991] env[61728]: DEBUG nova.virt.hardware [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 542.116175] env[61728]: DEBUG nova.virt.hardware [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 542.117381] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295a1804-2b1b-482d-a81a-95e18793af34 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.128140] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ed3620-9f85-4666-aa8f-bb1af90b736f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.161182] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.161182] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.167390] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a88423-44ef-c1e2-53e9-b59779e53a67, 'name': SearchDatastore_Task, 'duration_secs': 0.060753} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.167912] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-752c680b-3452-48fd-b146-ed4d85c97d1f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.174557] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 542.174557] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523e0270-cb8c-1cdb-8cf4-82fde95d7b3d" [ 542.174557] env[61728]: _type = "Task" [ 542.174557] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.183474] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523e0270-cb8c-1cdb-8cf4-82fde95d7b3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.194528] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463518, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.406735] env[61728]: DEBUG nova.network.neutron [req-96322bc4-93ad-4bfa-b168-adf9edeeedcb req-5ec928bb-0b57-48af-90a5-f105cc245be7 service nova] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Updated VIF entry in instance network info cache for port da4f38d4-527c-47c4-9fa7-a5009a0a4e6c. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 542.407119] env[61728]: DEBUG nova.network.neutron [req-96322bc4-93ad-4bfa-b168-adf9edeeedcb req-5ec928bb-0b57-48af-90a5-f105cc245be7 service nova] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Updating instance_info_cache with network_info: [{"id": "da4f38d4-527c-47c4-9fa7-a5009a0a4e6c", "address": "fa:16:3e:2e:c1:36", "network": {"id": "4da787ba-7f1e-46fb-9902-d90ec8854a65", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-249168995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f587ea8b8fe432aafd3e0ecba054bd0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48937bd1-23dc-413f-b46b-59bf9e709aa0", "external-id": "nsx-vlan-transportzone-160", "segmentation_id": 160, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda4f38d4-52", "ovs_interfaceid": "da4f38d4-527c-47c4-9fa7-a5009a0a4e6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.436894] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463519, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076029} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.437264] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 542.440178] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ec95fa-64d7-4847-ab88-091897c47f9a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.465941] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Reconfiguring VM instance instance-0000000c to attach disk [datastore1] e56a59a6-fc0f-45ca-810a-76b18bae173a/e56a59a6-fc0f-45ca-810a-76b18bae173a.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 542.468169] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa97a62c-694a-47d4-8e98-d26b089bd0c9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.487619] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Acquiring lock "0f75629b-bdec-4593-aceb-cd0478972ccc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.488170] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Lock "0f75629b-bdec-4593-aceb-cd0478972ccc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.498023] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 542.498023] env[61728]: value = "task-463523" [ 542.498023] env[61728]: _type = "Task" [ 542.498023] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.507023] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463523, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.561166] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463522, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.612182] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.570s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.612412] env[61728]: INFO nova.compute.manager [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Migrating [ 542.612774] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.613142] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.614314] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.714s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.615951] env[61728]: INFO nova.compute.claims [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 542.619457] env[61728]: INFO nova.compute.rpcapi [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 542.620063] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.688997] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523e0270-cb8c-1cdb-8cf4-82fde95d7b3d, 'name': SearchDatastore_Task, 'duration_secs': 0.012034} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.692862] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.693227] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618/ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 542.694332] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21fdc2c4-1ee6-4989-a26e-fbc727ef3856 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.709056] env[61728]: DEBUG oslo_vmware.api [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463518, 'name': PowerOnVM_Task, 'duration_secs': 0.811002} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.711158] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 542.712226] env[61728]: INFO nova.compute.manager [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Took 8.48 seconds to spawn the instance on the hypervisor. [ 542.712718] env[61728]: DEBUG nova.compute.manager [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 542.713034] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 542.713034] env[61728]: value = "task-463524" [ 542.713034] env[61728]: _type = "Task" [ 542.713034] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.714030] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d06526f-0a61-4ec0-830d-15041669e882 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.731343] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463524, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.849282] env[61728]: DEBUG nova.compute.manager [req-e7b7d679-d8cb-4972-9e10-5dd805cbad86 req-28f497de-0afd-4638-a8e0-dfc9d91727f3 service nova] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Received event network-vif-plugged-7af87a49-cf84-45ac-8bb4-e6d6062e4db0 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 542.849651] env[61728]: DEBUG oslo_concurrency.lockutils [req-e7b7d679-d8cb-4972-9e10-5dd805cbad86 req-28f497de-0afd-4638-a8e0-dfc9d91727f3 service nova] Acquiring lock "e56a59a6-fc0f-45ca-810a-76b18bae173a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.850753] env[61728]: DEBUG oslo_concurrency.lockutils [req-e7b7d679-d8cb-4972-9e10-5dd805cbad86 req-28f497de-0afd-4638-a8e0-dfc9d91727f3 service nova] Lock "e56a59a6-fc0f-45ca-810a-76b18bae173a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.851028] env[61728]: DEBUG oslo_concurrency.lockutils [req-e7b7d679-d8cb-4972-9e10-5dd805cbad86 req-28f497de-0afd-4638-a8e0-dfc9d91727f3 service nova] Lock "e56a59a6-fc0f-45ca-810a-76b18bae173a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.851160] env[61728]: DEBUG nova.compute.manager [req-e7b7d679-d8cb-4972-9e10-5dd805cbad86 req-28f497de-0afd-4638-a8e0-dfc9d91727f3 service nova] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] No waiting events found dispatching network-vif-plugged-7af87a49-cf84-45ac-8bb4-e6d6062e4db0 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 542.851349] env[61728]: WARNING nova.compute.manager [req-e7b7d679-d8cb-4972-9e10-5dd805cbad86 req-28f497de-0afd-4638-a8e0-dfc9d91727f3 service nova] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Received unexpected event network-vif-plugged-7af87a49-cf84-45ac-8bb4-e6d6062e4db0 for instance with vm_state building and task_state spawning. [ 542.851493] env[61728]: DEBUG nova.compute.manager [req-e7b7d679-d8cb-4972-9e10-5dd805cbad86 req-28f497de-0afd-4638-a8e0-dfc9d91727f3 service nova] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Received event network-changed-7af87a49-cf84-45ac-8bb4-e6d6062e4db0 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 542.851675] env[61728]: DEBUG nova.compute.manager [req-e7b7d679-d8cb-4972-9e10-5dd805cbad86 req-28f497de-0afd-4638-a8e0-dfc9d91727f3 service nova] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Refreshing instance network info cache due to event network-changed-7af87a49-cf84-45ac-8bb4-e6d6062e4db0. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 542.851939] env[61728]: DEBUG oslo_concurrency.lockutils [req-e7b7d679-d8cb-4972-9e10-5dd805cbad86 req-28f497de-0afd-4638-a8e0-dfc9d91727f3 service nova] Acquiring lock "refresh_cache-e56a59a6-fc0f-45ca-810a-76b18bae173a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.852049] env[61728]: DEBUG oslo_concurrency.lockutils [req-e7b7d679-d8cb-4972-9e10-5dd805cbad86 req-28f497de-0afd-4638-a8e0-dfc9d91727f3 service nova] Acquired lock "refresh_cache-e56a59a6-fc0f-45ca-810a-76b18bae173a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.852465] env[61728]: DEBUG nova.network.neutron [req-e7b7d679-d8cb-4972-9e10-5dd805cbad86 req-28f497de-0afd-4638-a8e0-dfc9d91727f3 service nova] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Refreshing network info cache for port 7af87a49-cf84-45ac-8bb4-e6d6062e4db0 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 542.909999] env[61728]: DEBUG oslo_concurrency.lockutils [req-96322bc4-93ad-4bfa-b168-adf9edeeedcb req-5ec928bb-0b57-48af-90a5-f105cc245be7 service nova] Releasing lock "refresh_cache-adb58292-62cd-4142-ad66-db014bcbdc49" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.009981] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463523, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.056783] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463522, 'name': CreateVM_Task, 'duration_secs': 0.516597} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.057025] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 543.058131] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.058304] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.059096] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 543.059245] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d31f455b-dbae-4e49-ae8b-3aee94b1ecdf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.065608] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 543.065608] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526f15a4-d9b6-2d58-e6c3-581b624be342" [ 543.065608] env[61728]: _type = "Task" [ 543.065608] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.075412] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526f15a4-d9b6-2d58-e6c3-581b624be342, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.098540] env[61728]: DEBUG nova.network.neutron [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Successfully updated port: edc90742-f02d-42ad-85cd-b4242f797c33 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 543.143408] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.143593] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.143657] env[61728]: DEBUG nova.network.neutron [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 543.230412] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463524, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511612} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.230787] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618/ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 543.230892] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 543.231232] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2e21d0ab-960f-4836-9220-481177c9059e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.245626] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 543.245626] env[61728]: value = "task-463525" [ 543.245626] env[61728]: _type = "Task" [ 543.245626] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.257619] env[61728]: INFO nova.compute.manager [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Took 26.97 seconds to build instance. [ 543.262592] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463525, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.486993] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Acquiring lock "fbd521c8-9618-4b5c-839e-50d8631fa7b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.486993] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Lock "fbd521c8-9618-4b5c-839e-50d8631fa7b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.509916] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463523, 'name': ReconfigVM_Task, 'duration_secs': 0.532829} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.510563] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Reconfigured VM instance instance-0000000c to attach disk [datastore1] e56a59a6-fc0f-45ca-810a-76b18bae173a/e56a59a6-fc0f-45ca-810a-76b18bae173a.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 543.510847] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-335dc849-8719-46c8-8f20-9921b59d5ffd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.519598] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 543.519598] env[61728]: value = "task-463526" [ 543.519598] env[61728]: _type = "Task" [ 543.519598] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.531118] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463526, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.582215] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526f15a4-d9b6-2d58-e6c3-581b624be342, 'name': SearchDatastore_Task, 'duration_secs': 0.047742} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.587957] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.587957] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 543.587957] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.587957] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.589180] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 543.589605] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2481f7fb-1be0-443f-9537-a9e7af1cb7c4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.601485] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.601654] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.601803] env[61728]: DEBUG nova.network.neutron [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 543.604242] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 543.604462] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 543.605564] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b826d61-1734-4b4f-8ff0-238af711aabe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.617237] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 543.617237] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5299d816-6902-4f4b-4017-4e422732b604" [ 543.617237] env[61728]: _type = "Task" [ 543.617237] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.630942] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5299d816-6902-4f4b-4017-4e422732b604, 'name': SearchDatastore_Task, 'duration_secs': 0.011507} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.633078] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce0d6324-97f6-4a5a-a061-082b8426f44d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.642993] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 543.642993] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e64c96-978f-dc39-abb7-d95086926aa4" [ 543.642993] env[61728]: _type = "Task" [ 543.642993] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.655688] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e64c96-978f-dc39-abb7-d95086926aa4, 'name': SearchDatastore_Task, 'duration_secs': 0.010832} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.655985] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.656680] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3c4f7d53-6ba6-4e2c-b324-e651640b304b/3c4f7d53-6ba6-4e2c-b324-e651640b304b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 543.656680] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aecd8383-828c-4a6d-8059-478db63faf9b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.665223] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 543.665223] env[61728]: value = "task-463527" [ 543.665223] env[61728]: _type = "Task" [ 543.665223] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.677907] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463527, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.761563] env[61728]: DEBUG oslo_concurrency.lockutils [None req-831866e5-ee8f-4723-b6f3-5de9e5aafe1a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "adb58292-62cd-4142-ad66-db014bcbdc49" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.086s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.761912] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463525, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.158982} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.763265] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 543.764400] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ffc11a-2a2d-4d15-b59c-9340f406c2d4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.797241] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Reconfiguring VM instance instance-00000008 to attach disk [datastore1] ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618/ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 543.798601] env[61728]: DEBUG nova.network.neutron [req-e7b7d679-d8cb-4972-9e10-5dd805cbad86 req-28f497de-0afd-4638-a8e0-dfc9d91727f3 service nova] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Updated VIF entry in instance network info cache for port 7af87a49-cf84-45ac-8bb4-e6d6062e4db0. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 543.800358] env[61728]: DEBUG nova.network.neutron [req-e7b7d679-d8cb-4972-9e10-5dd805cbad86 req-28f497de-0afd-4638-a8e0-dfc9d91727f3 service nova] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Updating instance_info_cache with network_info: [{"id": "7af87a49-cf84-45ac-8bb4-e6d6062e4db0", "address": "fa:16:3e:04:17:42", "network": {"id": "4da787ba-7f1e-46fb-9902-d90ec8854a65", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-249168995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f587ea8b8fe432aafd3e0ecba054bd0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48937bd1-23dc-413f-b46b-59bf9e709aa0", "external-id": "nsx-vlan-transportzone-160", "segmentation_id": 160, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7af87a49-cf", "ovs_interfaceid": "7af87a49-cf84-45ac-8bb4-e6d6062e4db0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.803397] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed8cfcf6-5389-4f27-828b-8247ce5c1ea8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.833848] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 543.833848] env[61728]: value = "task-463528" [ 543.833848] env[61728]: _type = "Task" [ 543.833848] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.849463] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463528, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.035769] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463526, 'name': Rename_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.150010] env[61728]: DEBUG nova.network.neutron [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.183125] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463527, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45554} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.186525] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3c4f7d53-6ba6-4e2c-b324-e651640b304b/3c4f7d53-6ba6-4e2c-b324-e651640b304b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 544.187257] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 544.189492] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b62b5c28-10da-4150-b484-ed7a16b6ccc9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.201927] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 544.201927] env[61728]: value = "task-463529" [ 544.201927] env[61728]: _type = "Task" [ 544.201927] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.214371] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463529, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.256134] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7e6001-1411-438c-88b8-db27ec3c9e9d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.264660] env[61728]: DEBUG nova.compute.manager [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 544.270245] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7682105e-1eab-4221-8d69-7586dc95b82e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.304110] env[61728]: DEBUG nova.network.neutron [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating instance_info_cache with network_info: [{"id": "1d776ec6-69e5-4f30-8201-13c69916cca0", "address": "fa:16:3e:6c:2d:56", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.159", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d776ec6-69", "ovs_interfaceid": "1d776ec6-69e5-4f30-8201-13c69916cca0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.306331] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e333a1-c339-41fd-b74b-36467b163362 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.314972] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d75932e5-d2e2-4480-a08f-321e4811f980 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.330409] env[61728]: DEBUG oslo_concurrency.lockutils [req-e7b7d679-d8cb-4972-9e10-5dd805cbad86 req-28f497de-0afd-4638-a8e0-dfc9d91727f3 service nova] Releasing lock "refresh_cache-e56a59a6-fc0f-45ca-810a-76b18bae173a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.332012] env[61728]: DEBUG nova.compute.provider_tree [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 544.346621] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463528, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.374437] env[61728]: DEBUG nova.network.neutron [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Updating instance_info_cache with network_info: [{"id": "edc90742-f02d-42ad-85cd-b4242f797c33", "address": "fa:16:3e:aa:75:a5", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedc90742-f0", "ovs_interfaceid": "edc90742-f02d-42ad-85cd-b4242f797c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.533043] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463526, 'name': Rename_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.621359] env[61728]: DEBUG nova.compute.manager [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Received event network-vif-plugged-132601da-9151-425d-afd3-3878b0424d7e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 544.621689] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] Acquiring lock "3c4f7d53-6ba6-4e2c-b324-e651640b304b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.621914] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] Lock "3c4f7d53-6ba6-4e2c-b324-e651640b304b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.622592] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] Lock "3c4f7d53-6ba6-4e2c-b324-e651640b304b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.622592] env[61728]: DEBUG nova.compute.manager [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] No waiting events found dispatching network-vif-plugged-132601da-9151-425d-afd3-3878b0424d7e {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 544.622592] env[61728]: WARNING nova.compute.manager [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Received unexpected event network-vif-plugged-132601da-9151-425d-afd3-3878b0424d7e for instance with vm_state building and task_state spawning. [ 544.622848] env[61728]: DEBUG nova.compute.manager [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Received event network-changed-132601da-9151-425d-afd3-3878b0424d7e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 544.622848] env[61728]: DEBUG nova.compute.manager [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Refreshing instance network info cache due to event network-changed-132601da-9151-425d-afd3-3878b0424d7e. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 544.622994] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] Acquiring lock "refresh_cache-3c4f7d53-6ba6-4e2c-b324-e651640b304b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.623129] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] Acquired lock "refresh_cache-3c4f7d53-6ba6-4e2c-b324-e651640b304b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.623286] env[61728]: DEBUG nova.network.neutron [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Refreshing network info cache for port 132601da-9151-425d-afd3-3878b0424d7e {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 544.717787] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463529, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.790485] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.809349] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Releasing lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.849898] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463528, 'name': ReconfigVM_Task, 'duration_secs': 0.525701} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.849898] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Reconfigured VM instance instance-00000008 to attach disk [datastore1] ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618/ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 544.850742] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4b740f69-a829-4a53-aa08-d2fd587b12ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.859473] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 544.859473] env[61728]: value = "task-463530" [ 544.859473] env[61728]: _type = "Task" [ 544.859473] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.865547] env[61728]: ERROR nova.scheduler.client.report [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [req-e74a4eb3-b181-4a23-81b1-cd0b08731dcb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e74a4eb3-b181-4a23-81b1-cd0b08731dcb"}]} [ 544.871199] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463530, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.877740] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.878091] env[61728]: DEBUG nova.compute.manager [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Instance network_info: |[{"id": "edc90742-f02d-42ad-85cd-b4242f797c33", "address": "fa:16:3e:aa:75:a5", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedc90742-f0", "ovs_interfaceid": "edc90742-f02d-42ad-85cd-b4242f797c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 544.878668] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:75:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'edc90742-f02d-42ad-85cd-b4242f797c33', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 544.886587] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Creating folder: Project (32bb35e1dfce40e48be08bb568d3f2b6). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 544.887920] env[61728]: DEBUG nova.scheduler.client.report [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 544.890267] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4ca574cd-c0b8-4c22-b26c-6ffbe356a908 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.902566] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Created folder: Project (32bb35e1dfce40e48be08bb568d3f2b6) in parent group-v121913. [ 544.902769] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Creating folder: Instances. Parent ref: group-v121956. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 544.903025] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f42909f-f439-410b-b168-5b24d3c77da3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.917782] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Created folder: Instances in parent group-v121956. [ 544.918041] env[61728]: DEBUG oslo.service.loopingcall [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 544.918568] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 544.918675] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-70c4143c-4f71-4599-b3c7-64da067b4d5c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.934065] env[61728]: DEBUG nova.scheduler.client.report [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 544.934297] env[61728]: DEBUG nova.compute.provider_tree [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 544.943921] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 544.943921] env[61728]: value = "task-463533" [ 544.943921] env[61728]: _type = "Task" [ 544.943921] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.948115] env[61728]: DEBUG nova.scheduler.client.report [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 544.958738] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463533, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.975410] env[61728]: DEBUG nova.scheduler.client.report [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 545.040160] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463526, 'name': Rename_Task, 'duration_secs': 1.22174} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.040365] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 545.040631] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3be140fc-f3b1-4025-b7ab-60bd7bf0bc1e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.050834] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 545.050834] env[61728]: value = "task-463534" [ 545.050834] env[61728]: _type = "Task" [ 545.050834] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.062894] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463534, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.217829] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463529, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.572811} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.218441] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 545.219966] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d53cc30-04f3-4ee0-b790-b14e733f5050 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.260715] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Reconfiguring VM instance instance-0000000d to attach disk [datastore1] 3c4f7d53-6ba6-4e2c-b324-e651640b304b/3c4f7d53-6ba6-4e2c-b324-e651640b304b.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 545.261707] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0211d93-f4eb-4e3b-af8a-31030ca328dd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.291805] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 545.291805] env[61728]: value = "task-463535" [ 545.291805] env[61728]: _type = "Task" [ 545.291805] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.306831] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463535, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.378169] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463530, 'name': Rename_Task, 'duration_secs': 0.240718} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.378169] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 545.378327] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61e32788-730d-4e50-9a32-43be4c4321fa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.389585] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 545.389585] env[61728]: value = "task-463536" [ 545.389585] env[61728]: _type = "Task" [ 545.389585] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.400233] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463536, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.454560] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463533, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.467559] env[61728]: DEBUG nova.network.neutron [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Updated VIF entry in instance network info cache for port 132601da-9151-425d-afd3-3878b0424d7e. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 545.467906] env[61728]: DEBUG nova.network.neutron [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Updating instance_info_cache with network_info: [{"id": "132601da-9151-425d-afd3-3878b0424d7e", "address": "fa:16:3e:94:a8:a2", "network": {"id": "17a8ea01-7320-4089-9138-a6fd17ac4054", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-599506352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "383f38592de14b4ba93fa9a6eeec40aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap132601da-91", "ovs_interfaceid": "132601da-9151-425d-afd3-3878b0424d7e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.565186] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463534, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.612926] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91766a4f-349f-4ea4-84e7-0b8763432d73 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.630903] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db3722a4-b227-4cd8-927a-635d2e7b7f5f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.682019] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a5212ed-4761-4f1f-8180-7db29abfb527 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.694229] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f58554a4-e1e3-4a77-aabd-ab34e16643c5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.700113] env[61728]: DEBUG oslo_concurrency.lockutils [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Acquiring lock "b86bf217-d1f8-4d37-aa87-3a2d27d70420" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.700396] env[61728]: DEBUG oslo_concurrency.lockutils [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Lock "b86bf217-d1f8-4d37-aa87-3a2d27d70420" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.700632] env[61728]: DEBUG oslo_concurrency.lockutils [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Acquiring lock "b86bf217-d1f8-4d37-aa87-3a2d27d70420-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.700816] env[61728]: DEBUG oslo_concurrency.lockutils [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Lock "b86bf217-d1f8-4d37-aa87-3a2d27d70420-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.701039] env[61728]: DEBUG oslo_concurrency.lockutils [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Lock "b86bf217-d1f8-4d37-aa87-3a2d27d70420-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.703903] env[61728]: INFO nova.compute.manager [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Terminating instance [ 545.705951] env[61728]: DEBUG nova.compute.manager [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 545.706204] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 545.707017] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e55b901c-f96d-4a31-828c-bbf3b39427bf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.721039] env[61728]: DEBUG nova.compute.provider_tree [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 545.728752] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 545.728931] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff6a495f-7589-4a06-baf1-92a3546f5fc0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.737637] env[61728]: DEBUG oslo_vmware.api [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Waiting for the task: (returnval){ [ 545.737637] env[61728]: value = "task-463537" [ 545.737637] env[61728]: _type = "Task" [ 545.737637] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.748498] env[61728]: DEBUG oslo_vmware.api [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463537, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.803837] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463535, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.901181] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463536, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.961095] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463533, 'name': CreateVM_Task, 'duration_secs': 0.607961} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.961455] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 545.962705] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.962705] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.962923] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 545.965216] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8c03e57-cd25-4ba9-ad4c-2a9a409a2c58 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.971032] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] Releasing lock "refresh_cache-3c4f7d53-6ba6-4e2c-b324-e651640b304b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.971032] env[61728]: DEBUG nova.compute.manager [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Received event network-vif-plugged-edc90742-f02d-42ad-85cd-b4242f797c33 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 545.971260] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] Acquiring lock "620bb472-c36e-4c56-acdf-42e02614856b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.971407] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] Lock "620bb472-c36e-4c56-acdf-42e02614856b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.971725] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] Lock "620bb472-c36e-4c56-acdf-42e02614856b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.972093] env[61728]: DEBUG nova.compute.manager [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] No waiting events found dispatching network-vif-plugged-edc90742-f02d-42ad-85cd-b4242f797c33 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 545.972346] env[61728]: WARNING nova.compute.manager [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Received unexpected event network-vif-plugged-edc90742-f02d-42ad-85cd-b4242f797c33 for instance with vm_state building and task_state spawning. [ 545.972789] env[61728]: DEBUG nova.compute.manager [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Received event network-changed-edc90742-f02d-42ad-85cd-b4242f797c33 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 545.973257] env[61728]: DEBUG nova.compute.manager [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Refreshing instance network info cache due to event network-changed-edc90742-f02d-42ad-85cd-b4242f797c33. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 545.973441] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] Acquiring lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.973647] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] Acquired lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.973858] env[61728]: DEBUG nova.network.neutron [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Refreshing network info cache for port edc90742-f02d-42ad-85cd-b4242f797c33 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 545.975245] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 545.975245] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52962a53-fb6e-076f-4360-a12ccf749f88" [ 545.975245] env[61728]: _type = "Task" [ 545.975245] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.988130] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52962a53-fb6e-076f-4360-a12ccf749f88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.067790] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463534, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.249243] env[61728]: DEBUG oslo_vmware.api [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463537, 'name': PowerOffVM_Task, 'duration_secs': 0.394262} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.249560] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 546.249696] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 546.249942] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f15f5aa-666d-44b8-9860-3778f900552f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.255921] env[61728]: DEBUG nova.scheduler.client.report [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 39 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 546.256217] env[61728]: DEBUG nova.compute.provider_tree [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 39 to 40 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 546.256217] env[61728]: DEBUG nova.compute.provider_tree [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 546.303849] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463535, 'name': ReconfigVM_Task, 'duration_secs': 0.555537} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.304214] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Reconfigured VM instance instance-0000000d to attach disk [datastore1] 3c4f7d53-6ba6-4e2c-b324-e651640b304b/3c4f7d53-6ba6-4e2c-b324-e651640b304b.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 546.304899] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e5bc9b6-8ccf-4e15-a9f1-9f5f753abfaa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.314091] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 546.314091] env[61728]: value = "task-463539" [ 546.314091] env[61728]: _type = "Task" [ 546.314091] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.331348] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 546.331348] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 546.331348] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Deleting the datastore file [datastore1] b86bf217-d1f8-4d37-aa87-3a2d27d70420 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 546.331685] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463539, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.332350] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f50da12-37c3-4305-9c17-cae542516632 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.334782] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ebb5c9d-6f61-43b4-b061-b630d3c0df0c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.356271] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating instance '9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c' progress to 0 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 546.362259] env[61728]: DEBUG oslo_vmware.api [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Waiting for the task: (returnval){ [ 546.362259] env[61728]: value = "task-463540" [ 546.362259] env[61728]: _type = "Task" [ 546.362259] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.372105] env[61728]: DEBUG oslo_vmware.api [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463540, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.401873] env[61728]: DEBUG oslo_vmware.api [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463536, 'name': PowerOnVM_Task, 'duration_secs': 0.592955} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.401873] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 546.401873] env[61728]: DEBUG nova.compute.manager [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 546.402160] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7016752-9cc2-4f3a-87e8-36f36815c8f3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.494026] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52962a53-fb6e-076f-4360-a12ccf749f88, 'name': SearchDatastore_Task, 'duration_secs': 0.015472} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.494026] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.494026] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 546.494026] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.494468] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.494468] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 546.494468] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-300bcc70-ee6c-403e-965d-cc8ede3d7270 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.506070] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 546.506070] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 546.506070] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6eea8c5-d233-4070-b36d-c1895fea7e5a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.513158] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 546.513158] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5289e5b2-0e7d-909f-d0e4-2bf48f219947" [ 546.513158] env[61728]: _type = "Task" [ 546.513158] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.524563] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5289e5b2-0e7d-909f-d0e4-2bf48f219947, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.566342] env[61728]: DEBUG oslo_vmware.api [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463534, 'name': PowerOnVM_Task, 'duration_secs': 1.184462} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.566705] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 546.566836] env[61728]: INFO nova.compute.manager [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Took 9.73 seconds to spawn the instance on the hypervisor. [ 546.567026] env[61728]: DEBUG nova.compute.manager [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 546.567785] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecffdd6c-de3e-47ae-a658-bdc43f59f860 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.767677] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.153s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.768444] env[61728]: DEBUG nova.compute.manager [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 546.772272] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.622s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.776130] env[61728]: INFO nova.compute.claims [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 546.828534] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463539, 'name': Rename_Task, 'duration_secs': 0.222339} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.828831] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 546.829089] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f956893-d7ab-4767-a763-0230460626e1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.842764] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 546.842764] env[61728]: value = "task-463541" [ 546.842764] env[61728]: _type = "Task" [ 546.842764] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.853601] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463541, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.865374] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 546.865374] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28b36b07-fb50-47a6-88e7-3fdb42cc2f5c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.887905] env[61728]: DEBUG oslo_vmware.api [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Task: {'id': task-463540, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.306956} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.889427] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 546.889658] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 546.890301] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 546.890301] env[61728]: INFO nova.compute.manager [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Took 1.18 seconds to destroy the instance on the hypervisor. [ 546.890301] env[61728]: DEBUG oslo.service.loopingcall [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 546.890557] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 546.890557] env[61728]: value = "task-463542" [ 546.890557] env[61728]: _type = "Task" [ 546.890557] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.890774] env[61728]: DEBUG nova.compute.manager [-] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 546.890774] env[61728]: DEBUG nova.network.neutron [-] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 546.910726] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463542, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.911698] env[61728]: DEBUG nova.network.neutron [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Updated VIF entry in instance network info cache for port edc90742-f02d-42ad-85cd-b4242f797c33. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 546.912837] env[61728]: DEBUG nova.network.neutron [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Updating instance_info_cache with network_info: [{"id": "edc90742-f02d-42ad-85cd-b4242f797c33", "address": "fa:16:3e:aa:75:a5", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedc90742-f0", "ovs_interfaceid": "edc90742-f02d-42ad-85cd-b4242f797c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.930168] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.033768] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5289e5b2-0e7d-909f-d0e4-2bf48f219947, 'name': SearchDatastore_Task, 'duration_secs': 0.015739} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.035714] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-326eb90d-1247-4cc1-95d8-ccd5d52329f9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.046024] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 547.046024] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520b8760-01d3-b085-2e0e-39e12a1748a3" [ 547.046024] env[61728]: _type = "Task" [ 547.046024] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.056273] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520b8760-01d3-b085-2e0e-39e12a1748a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.095796] env[61728]: INFO nova.compute.manager [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Took 29.08 seconds to build instance. [ 547.287606] env[61728]: DEBUG nova.compute.utils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 547.289460] env[61728]: DEBUG nova.compute.manager [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 547.289703] env[61728]: DEBUG nova.network.neutron [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 547.354605] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463541, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.404367] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463542, 'name': PowerOffVM_Task, 'duration_secs': 0.253874} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.406066] env[61728]: DEBUG nova.policy [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31c50806197540cf85b9da7495bfe9e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9e4f01166b049579fe2527d66e4f768', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 547.407665] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 547.407914] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating instance '9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c' progress to 17 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 547.423857] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f548bd0-fc9f-4045-89a7-ba4f93a1e71d req-63a0cb6d-08d9-4326-96ae-30b2d6f00f64 service nova] Releasing lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.557027] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520b8760-01d3-b085-2e0e-39e12a1748a3, 'name': SearchDatastore_Task, 'duration_secs': 0.013599} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.557317] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.557564] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 620bb472-c36e-4c56-acdf-42e02614856b/620bb472-c36e-4c56-acdf-42e02614856b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 547.557856] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3e22dba1-6a00-400d-a04a-806689d8319b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.567088] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 547.567088] env[61728]: value = "task-463543" [ 547.567088] env[61728]: _type = "Task" [ 547.567088] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.576251] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463543, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.596383] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec2574df-e031-47dd-bdaa-67c62fbeed64 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "e56a59a6-fc0f-45ca-810a-76b18bae173a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.797s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.793172] env[61728]: DEBUG nova.compute.manager [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 547.861606] env[61728]: DEBUG oslo_vmware.api [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463541, 'name': PowerOnVM_Task, 'duration_secs': 0.850317} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.862387] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 547.862620] env[61728]: INFO nova.compute.manager [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Took 8.43 seconds to spawn the instance on the hypervisor. [ 547.862946] env[61728]: DEBUG nova.compute.manager [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 547.863716] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef06890-e0a8-4a6d-baf1-f1ff93b2c5c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.922573] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 547.922573] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 547.922573] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 547.922573] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 547.923777] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 547.923777] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 547.923777] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 547.923777] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 547.923777] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 547.924099] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 547.924099] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 547.934285] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-640efa8f-9fa1-4f29-acff-326202227962 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.955617] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 547.955617] env[61728]: value = "task-463544" [ 547.955617] env[61728]: _type = "Task" [ 547.955617] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.961177] env[61728]: DEBUG nova.network.neutron [-] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.970171] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463544, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.080317] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463543, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.101374] env[61728]: DEBUG nova.compute.manager [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.150439] env[61728]: DEBUG nova.compute.manager [req-07888e62-e146-49e4-bbc2-8ef031912687 req-6a0f7971-41f7-4462-b919-173f6540788c service nova] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Received event network-vif-deleted-94e809f2-e1ed-45f6-872f-ec7c798028aa {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 548.402342] env[61728]: INFO nova.compute.manager [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Took 30.37 seconds to build instance. [ 548.413976] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e97817-6762-4355-965f-776ceb25efb1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.425743] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c115dfab-8df0-4a42-a7d5-245b4e74ce27 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.464413] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd576a5-5f0a-4713-9e2a-995d524597c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.467253] env[61728]: INFO nova.compute.manager [-] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Took 1.58 seconds to deallocate network for instance. [ 548.481337] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463544, 'name': ReconfigVM_Task, 'duration_secs': 0.323285} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.482872] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ae0abc-a0e1-453c-9691-25c5a56e865a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.487085] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating instance '9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c' progress to 33 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 548.503064] env[61728]: DEBUG nova.compute.provider_tree [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 548.579628] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463543, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.599352} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.579959] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 620bb472-c36e-4c56-acdf-42e02614856b/620bb472-c36e-4c56-acdf-42e02614856b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 548.580154] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 548.580423] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7462ab25-b9d2-4519-944f-f0b3f4c63c7c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.588566] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 548.588566] env[61728]: value = "task-463545" [ 548.588566] env[61728]: _type = "Task" [ 548.588566] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.598796] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463545, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.628298] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.648366] env[61728]: DEBUG nova.network.neutron [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Successfully created port: f77e3955-f8e6-43aa-ad48-d8bc4d078422 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.805836] env[61728]: DEBUG nova.compute.manager [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 548.846478] env[61728]: DEBUG nova.virt.hardware [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.846869] env[61728]: DEBUG nova.virt.hardware [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.847039] env[61728]: DEBUG nova.virt.hardware [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.851087] env[61728]: DEBUG nova.virt.hardware [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.851770] env[61728]: DEBUG nova.virt.hardware [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.851770] env[61728]: DEBUG nova.virt.hardware [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.852175] env[61728]: DEBUG nova.virt.hardware [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.852307] env[61728]: DEBUG nova.virt.hardware [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.852439] env[61728]: DEBUG nova.virt.hardware [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.852651] env[61728]: DEBUG nova.virt.hardware [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.852812] env[61728]: DEBUG nova.virt.hardware [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.853987] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2859c5-2b53-4bb0-b67e-1d3642babbf8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.865078] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a3d492-b67e-488a-8cef-d3f03d9f4c60 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.904929] env[61728]: DEBUG oslo_concurrency.lockutils [None req-123bad1f-1fba-4abc-92bd-c016b2efd362 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "3c4f7d53-6ba6-4e2c-b324-e651640b304b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.248s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.983105] env[61728]: DEBUG oslo_concurrency.lockutils [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.993618] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.993862] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.994026] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.994206] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.994359] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.994507] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.994812] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.995671] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.995764] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.996477] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.996477] env[61728]: DEBUG nova.virt.hardware [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 549.001714] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Reconfiguring VM instance instance-00000001 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 549.003064] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-117f930d-9759-40b5-aa6a-4b0732da3935 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.029625] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 549.029625] env[61728]: value = "task-463546" [ 549.029625] env[61728]: _type = "Task" [ 549.029625] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.039358] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463546, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.065765] env[61728]: DEBUG nova.scheduler.client.report [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 40 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 549.066055] env[61728]: DEBUG nova.compute.provider_tree [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 40 to 41 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 549.066239] env[61728]: DEBUG nova.compute.provider_tree [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 549.108428] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463545, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087255} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.108729] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 549.109583] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a293a05-0034-4c52-9d3a-7512a5b568ea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.147743] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Reconfiguring VM instance instance-0000000e to attach disk [datastore1] 620bb472-c36e-4c56-acdf-42e02614856b/620bb472-c36e-4c56-acdf-42e02614856b.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 549.148094] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9684c5ac-b009-4e79-9bbc-8fc6325afcf9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.173123] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 549.173123] env[61728]: value = "task-463547" [ 549.173123] env[61728]: _type = "Task" [ 549.173123] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.183724] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463547, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.412013] env[61728]: DEBUG nova.compute.manager [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 549.544383] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463546, 'name': ReconfigVM_Task, 'duration_secs': 0.476682} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.546701] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Reconfigured VM instance instance-00000001 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 549.547821] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95cbe052-3bab-4d96-aa99-2db0dc370eda {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.574161] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c/9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 549.575476] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.804s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.575956] env[61728]: DEBUG nova.compute.manager [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 549.578460] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d486bf9c-6814-4398-a0e8-87f8b56a591b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.592826] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.469s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.593138] env[61728]: DEBUG nova.objects.instance [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Lazy-loading 'resources' on Instance uuid d75ab8eb-f15c-4e0f-8376-6d91aa26ab58 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 549.601892] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 549.601892] env[61728]: value = "task-463548" [ 549.601892] env[61728]: _type = "Task" [ 549.601892] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.612719] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463548, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.687570] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463547, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.943465] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.094265] env[61728]: DEBUG nova.compute.utils [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 550.095637] env[61728]: DEBUG nova.compute.manager [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Not allocating networking since 'none' was specified. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 550.120883] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.189121] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463547, 'name': ReconfigVM_Task, 'duration_secs': 0.678421} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.191879] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Reconfigured VM instance instance-0000000e to attach disk [datastore1] 620bb472-c36e-4c56-acdf-42e02614856b/620bb472-c36e-4c56-acdf-42e02614856b.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 550.192720] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d5aec15-6e67-4ab2-b796-f6c83c813a5e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.202156] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 550.202156] env[61728]: value = "task-463549" [ 550.202156] env[61728]: _type = "Task" [ 550.202156] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.219046] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463549, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.589554] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.589974] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.003s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.590374] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.591310] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.591606] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.594587] env[61728]: INFO nova.compute.manager [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Terminating instance [ 550.601367] env[61728]: DEBUG nova.compute.manager [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 550.601688] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 550.602968] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc92599a-8950-4099-a05e-16b021389460 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.609836] env[61728]: DEBUG nova.compute.manager [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 550.628613] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463548, 'name': ReconfigVM_Task, 'duration_secs': 0.584303} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.631008] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Reconfigured VM instance instance-00000001 to attach disk [datastore1] 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c/9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 550.631301] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating instance '9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c' progress to 50 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 550.637328] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 550.641029] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54e08ee2-49e9-408b-bff6-5d8f58645702 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.650906] env[61728]: DEBUG oslo_vmware.api [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 550.650906] env[61728]: value = "task-463550" [ 550.650906] env[61728]: _type = "Task" [ 550.650906] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.660105] env[61728]: DEBUG oslo_vmware.api [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463550, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.719191] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463549, 'name': Rename_Task, 'duration_secs': 0.190271} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.720392] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d45037f-744c-4740-8d47-7a34151f3c01 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.723756] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 550.724360] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c9a1a1bc-f3fd-4061-a687-4e420db53327 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.733249] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6c48f8-c696-49d1-88ce-524bf61c1145 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.738483] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 550.738483] env[61728]: value = "task-463551" [ 550.738483] env[61728]: _type = "Task" [ 550.738483] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.774289] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e545706e-a3f0-4479-9e5d-799e15071806 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.781205] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463551, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.788671] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92bcf7bc-adaf-4e1e-84f7-1fc3e11157df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.806798] env[61728]: DEBUG nova.compute.provider_tree [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.143047] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba045991-652d-4a18-bb57-80c08c6aba65 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.179679] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d1e5526-bc41-48c7-bb53-996a21a9f20b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.205189] env[61728]: DEBUG oslo_vmware.api [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463550, 'name': PowerOffVM_Task, 'duration_secs': 0.277967} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.207485] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating instance '9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c' progress to 67 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 551.209019] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 551.209195] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 551.209896] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a3747a8-8c3c-48a6-8359-1ae9883298df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.251595] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463551, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.297558] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 551.297558] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 551.297728] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleting the datastore file [datastore1] ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 551.298963] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-73a8e2f6-8cea-4aa2-9081-8fc24f3b7c9c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.307582] env[61728]: DEBUG oslo_vmware.api [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 551.307582] env[61728]: value = "task-463553" [ 551.307582] env[61728]: _type = "Task" [ 551.307582] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.312819] env[61728]: DEBUG nova.scheduler.client.report [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 551.325743] env[61728]: DEBUG oslo_vmware.api [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463553, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.352037] env[61728]: DEBUG nova.network.neutron [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Successfully updated port: f77e3955-f8e6-43aa-ad48-d8bc4d078422 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 551.625299] env[61728]: DEBUG nova.compute.manager [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 551.637301] env[61728]: DEBUG nova.compute.manager [req-12c3b252-c0d3-4155-867e-46c995795dee req-4c01695f-64c5-400c-b2fc-4e21dfab9665 service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Received event network-vif-plugged-f77e3955-f8e6-43aa-ad48-d8bc4d078422 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 551.637707] env[61728]: DEBUG oslo_concurrency.lockutils [req-12c3b252-c0d3-4155-867e-46c995795dee req-4c01695f-64c5-400c-b2fc-4e21dfab9665 service nova] Acquiring lock "8c4b7227-1a01-4400-beb8-2df27e17e329-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.639592] env[61728]: DEBUG oslo_concurrency.lockutils [req-12c3b252-c0d3-4155-867e-46c995795dee req-4c01695f-64c5-400c-b2fc-4e21dfab9665 service nova] Lock "8c4b7227-1a01-4400-beb8-2df27e17e329-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.639592] env[61728]: DEBUG oslo_concurrency.lockutils [req-12c3b252-c0d3-4155-867e-46c995795dee req-4c01695f-64c5-400c-b2fc-4e21dfab9665 service nova] Lock "8c4b7227-1a01-4400-beb8-2df27e17e329-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.639592] env[61728]: DEBUG nova.compute.manager [req-12c3b252-c0d3-4155-867e-46c995795dee req-4c01695f-64c5-400c-b2fc-4e21dfab9665 service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] No waiting events found dispatching network-vif-plugged-f77e3955-f8e6-43aa-ad48-d8bc4d078422 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 551.639592] env[61728]: WARNING nova.compute.manager [req-12c3b252-c0d3-4155-867e-46c995795dee req-4c01695f-64c5-400c-b2fc-4e21dfab9665 service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Received unexpected event network-vif-plugged-f77e3955-f8e6-43aa-ad48-d8bc4d078422 for instance with vm_state building and task_state spawning. [ 551.666271] env[61728]: DEBUG nova.virt.hardware [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 551.666731] env[61728]: DEBUG nova.virt.hardware [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 551.666894] env[61728]: DEBUG nova.virt.hardware [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 551.667087] env[61728]: DEBUG nova.virt.hardware [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 551.667244] env[61728]: DEBUG nova.virt.hardware [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 551.667431] env[61728]: DEBUG nova.virt.hardware [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 551.667607] env[61728]: DEBUG nova.virt.hardware [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 551.667845] env[61728]: DEBUG nova.virt.hardware [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 551.668039] env[61728]: DEBUG nova.virt.hardware [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 551.668223] env[61728]: DEBUG nova.virt.hardware [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 551.668396] env[61728]: DEBUG nova.virt.hardware [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 551.669358] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d3d8e1-202f-479b-9037-dd23df8a2d27 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.681366] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b610460b-1de9-4e42-a8d2-81cfa2b871ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.702129] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Instance VIF info [] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 551.708301] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Creating folder: Project (73aa0b16695f40fcbe5df30b3ebfa994). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 551.709651] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b1da8ba5-ef77-4003-b5fa-a709e01b708a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.727632] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Created folder: Project (73aa0b16695f40fcbe5df30b3ebfa994) in parent group-v121913. [ 551.727892] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Creating folder: Instances. Parent ref: group-v121959. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 551.728227] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8756e784-407a-44fe-9e99-cd84f5f12abe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.740015] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Created folder: Instances in parent group-v121959. [ 551.740550] env[61728]: DEBUG oslo.service.loopingcall [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 551.740689] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 551.744865] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-df77fcd4-a9b4-41e5-ad1c-03e4f41dd89c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.762718] env[61728]: DEBUG oslo_vmware.api [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463551, 'name': PowerOnVM_Task, 'duration_secs': 0.677872} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.763873] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 551.764672] env[61728]: INFO nova.compute.manager [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Took 9.68 seconds to spawn the instance on the hypervisor. [ 551.764672] env[61728]: DEBUG nova.compute.manager [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 551.764672] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 551.764672] env[61728]: value = "task-463556" [ 551.764672] env[61728]: _type = "Task" [ 551.764672] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.765228] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d59c52f-7a31-41df-96fa-16b770ab44f8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.776555] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463556, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.805197] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "1ef96606-d960-4283-b372-e2bad47f9367" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.805772] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "1ef96606-d960-4283-b372-e2bad47f9367" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.819099] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.226s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.821175] env[61728]: DEBUG oslo_vmware.api [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463553, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.335779} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.822012] env[61728]: DEBUG nova.network.neutron [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Port 1d776ec6-69e5-4f30-8201-13c69916cca0 binding to destination host cpu-1 is already ACTIVE {{(pid=61728) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 551.823520] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.986s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.823730] env[61728]: DEBUG nova.objects.instance [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Lazy-loading 'resources' on Instance uuid 3cd416c6-9926-45ee-afe3-d3468e4e06c2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 551.825122] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 551.826029] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 551.826029] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 551.826029] env[61728]: INFO nova.compute.manager [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Took 1.22 seconds to destroy the instance on the hypervisor. [ 551.826029] env[61728]: DEBUG oslo.service.loopingcall [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 551.826950] env[61728]: DEBUG nova.compute.manager [-] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 551.826950] env[61728]: DEBUG nova.network.neutron [-] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 551.855595] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Acquiring lock "refresh_cache-8c4b7227-1a01-4400-beb8-2df27e17e329" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.855595] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Acquired lock "refresh_cache-8c4b7227-1a01-4400-beb8-2df27e17e329" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.855869] env[61728]: DEBUG nova.network.neutron [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 551.858634] env[61728]: INFO nova.scheduler.client.report [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Deleted allocations for instance d75ab8eb-f15c-4e0f-8376-6d91aa26ab58 [ 552.278962] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463556, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.295156] env[61728]: INFO nova.compute.manager [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Took 32.19 seconds to build instance. [ 552.372131] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f91b36bd-59b2-4b79-8085-127145b7e17b tempest-DeleteServersAdminTestJSON-1701885572 tempest-DeleteServersAdminTestJSON-1701885572-project-admin] Lock "d75ab8eb-f15c-4e0f-8376-6d91aa26ab58" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.641s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.428618] env[61728]: DEBUG nova.network.neutron [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.772023] env[61728]: DEBUG nova.network.neutron [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Updating instance_info_cache with network_info: [{"id": "f77e3955-f8e6-43aa-ad48-d8bc4d078422", "address": "fa:16:3e:af:89:5c", "network": {"id": "77ffa432-b9ec-4b9f-b25b-b7d2366d3b37", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-458592337-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e4f01166b049579fe2527d66e4f768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf77e3955-f8", "ovs_interfaceid": "f77e3955-f8e6-43aa-ad48-d8bc4d078422", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.783810] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463556, 'name': CreateVM_Task, 'duration_secs': 0.672443} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.784023] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 552.785018] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.785018] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.785018] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 552.785229] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abfd513d-b81d-4c15-8c15-ab6ab6957d2e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.791211] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 552.791211] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5246c637-205a-d3b9-e012-fb07ace142ed" [ 552.791211] env[61728]: _type = "Task" [ 552.791211] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.801036] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7b058c9e-d032-42a2-80d5-ed5642a6843e tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "620bb472-c36e-4c56-acdf-42e02614856b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.479s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.809600] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5246c637-205a-d3b9-e012-fb07ace142ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.856280] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.856280] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.856280] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.867597] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7695c8b3-872a-463d-b180-7caac3b572d8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.880128] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fdf7ddb-a717-4fe5-b91c-b1ae1bd74454 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.912519] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c94d1134-0ea8-4d88-a8a9-931fb047fe0c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.921909] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506cf307-66e7-4888-a3d9-c5c1edb39dbe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.936429] env[61728]: DEBUG nova.compute.provider_tree [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.277711] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Releasing lock "refresh_cache-8c4b7227-1a01-4400-beb8-2df27e17e329" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.281149] env[61728]: DEBUG nova.compute.manager [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Instance network_info: |[{"id": "f77e3955-f8e6-43aa-ad48-d8bc4d078422", "address": "fa:16:3e:af:89:5c", "network": {"id": "77ffa432-b9ec-4b9f-b25b-b7d2366d3b37", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-458592337-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e4f01166b049579fe2527d66e4f768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf77e3955-f8", "ovs_interfaceid": "f77e3955-f8e6-43aa-ad48-d8bc4d078422", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 553.281284] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:af:89:5c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78e1ebb0-0130-446b-bf73-a0e59bbb95cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f77e3955-f8e6-43aa-ad48-d8bc4d078422', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 553.287145] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Creating folder: Project (a9e4f01166b049579fe2527d66e4f768). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 553.287468] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7353e1a0-23f5-4a08-b124-973a3e79693d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.302561] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5246c637-205a-d3b9-e012-fb07ace142ed, 'name': SearchDatastore_Task, 'duration_secs': 0.025497} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.304372] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.304638] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 553.304878] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.305282] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.305282] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 553.305711] env[61728]: DEBUG nova.compute.manager [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 553.309287] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Created folder: Project (a9e4f01166b049579fe2527d66e4f768) in parent group-v121913. [ 553.310035] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Creating folder: Instances. Parent ref: group-v121962. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 553.310281] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-38897980-768a-4156-ba3b-806e01e7eba4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.313419] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-591fe64f-19b7-4165-9ab0-4392dc17e6d3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.326597] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Created folder: Instances in parent group-v121962. [ 553.326789] env[61728]: DEBUG oslo.service.loopingcall [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 553.329117] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 553.329117] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 553.329117] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 553.329723] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-891b5688-aa49-4c8c-b39c-d77610677a94 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.350673] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3706be88-fa60-4fcd-9323-89b3bde0f11b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.352839] env[61728]: DEBUG nova.network.neutron [-] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.360417] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 553.360417] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5239f2f7-838b-064f-e5e0-7bd89c5b3b02" [ 553.360417] env[61728]: _type = "Task" [ 553.360417] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.363608] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 553.363608] env[61728]: value = "task-463559" [ 553.363608] env[61728]: _type = "Task" [ 553.363608] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.380744] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463559, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.381059] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5239f2f7-838b-064f-e5e0-7bd89c5b3b02, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.439798] env[61728]: DEBUG nova.scheduler.client.report [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.586927] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquiring lock "7a60fcd0-01b0-40bc-93e7-6f5ce03232d4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.587445] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Lock "7a60fcd0-01b0-40bc-93e7-6f5ce03232d4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.839914] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.857843] env[61728]: INFO nova.compute.manager [-] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Took 2.03 seconds to deallocate network for instance. [ 553.881443] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5239f2f7-838b-064f-e5e0-7bd89c5b3b02, 'name': SearchDatastore_Task, 'duration_secs': 0.018844} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.885985] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463559, 'name': CreateVM_Task, 'duration_secs': 0.38192} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.886131] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91c34836-39d1-4a41-a70e-aceb86d121ee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.888898] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 553.891415] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.891415] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.891415] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 553.891415] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d59ede8-f9d8-435b-8ab3-e9df82b10580 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.897359] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 553.897359] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524c5522-8cc0-a318-57dc-f9aa79602b83" [ 553.897359] env[61728]: _type = "Task" [ 553.897359] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.898629] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Waiting for the task: (returnval){ [ 553.898629] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522c971e-9ac0-a11f-40ec-1fd7df69756e" [ 553.898629] env[61728]: _type = "Task" [ 553.898629] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.916147] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524c5522-8cc0-a318-57dc-f9aa79602b83, 'name': SearchDatastore_Task, 'duration_secs': 0.012003} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.919322] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.919835] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2/a0831461-ece1-43ee-92f6-34d7d4e673e2.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 553.920298] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522c971e-9ac0-a11f-40ec-1fd7df69756e, 'name': SearchDatastore_Task, 'duration_secs': 0.010373} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.920604] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-04dced72-d3c9-49a1-84f5-b0e6d604de5b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.924017] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.924017] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 553.924017] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.924299] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.924570] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 553.924928] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e58e6fe-1272-46c0-a9f8-b549596f344a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.936036] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 553.936036] env[61728]: value = "task-463560" [ 553.936036] env[61728]: _type = "Task" [ 553.936036] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.938019] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 553.938019] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 553.941239] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f8af57b-7d93-4d22-b547-f01ced4a5018 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.945957] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.122s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.953356] env[61728]: DEBUG oslo_concurrency.lockutils [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.581s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.954844] env[61728]: INFO nova.compute.claims [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 553.957932] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463560, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.959103] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Waiting for the task: (returnval){ [ 553.959103] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5296482b-70a9-d083-19b9-b1acb852db1f" [ 553.959103] env[61728]: _type = "Task" [ 553.959103] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.974795] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5296482b-70a9-d083-19b9-b1acb852db1f, 'name': SearchDatastore_Task, 'duration_secs': 0.013374} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.974795] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a176dc6-3fdb-4aa6-8f99-b444049f0b3e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.979235] env[61728]: INFO nova.scheduler.client.report [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Deleted allocations for instance 3cd416c6-9926-45ee-afe3-d3468e4e06c2 [ 553.982040] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Waiting for the task: (returnval){ [ 553.982040] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bcc09f-3f72-0a13-be35-3db40bf18fab" [ 553.982040] env[61728]: _type = "Task" [ 553.982040] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.002033] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bcc09f-3f72-0a13-be35-3db40bf18fab, 'name': SearchDatastore_Task, 'duration_secs': 0.011772} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.002309] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.002582] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 8c4b7227-1a01-4400-beb8-2df27e17e329/8c4b7227-1a01-4400-beb8-2df27e17e329.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 554.002856] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e21cd972-0bbe-4c45-b22b-8b2790c703e8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.015820] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Waiting for the task: (returnval){ [ 554.015820] env[61728]: value = "task-463561" [ 554.015820] env[61728]: _type = "Task" [ 554.015820] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.028594] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-463561, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.040943] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.041830] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.041830] env[61728]: DEBUG nova.network.neutron [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 554.368025] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.451318] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463560, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.496421] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d6b65be6-af01-46b0-85f6-dbf27b6710ec tempest-ServerDiagnosticsTest-530299411 tempest-ServerDiagnosticsTest-530299411-project-member] Lock "3cd416c6-9926-45ee-afe3-d3468e4e06c2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.052s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.527960] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-463561, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.660467] env[61728]: DEBUG nova.compute.manager [req-f416ec17-fa59-4327-b94e-a1a408571a7b req-5eac146d-1387-451f-8e77-7dd42391f4b2 service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Received event network-changed-f77e3955-f8e6-43aa-ad48-d8bc4d078422 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 554.660467] env[61728]: DEBUG nova.compute.manager [req-f416ec17-fa59-4327-b94e-a1a408571a7b req-5eac146d-1387-451f-8e77-7dd42391f4b2 service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Refreshing instance network info cache due to event network-changed-f77e3955-f8e6-43aa-ad48-d8bc4d078422. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 554.660828] env[61728]: DEBUG oslo_concurrency.lockutils [req-f416ec17-fa59-4327-b94e-a1a408571a7b req-5eac146d-1387-451f-8e77-7dd42391f4b2 service nova] Acquiring lock "refresh_cache-8c4b7227-1a01-4400-beb8-2df27e17e329" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.660932] env[61728]: DEBUG oslo_concurrency.lockutils [req-f416ec17-fa59-4327-b94e-a1a408571a7b req-5eac146d-1387-451f-8e77-7dd42391f4b2 service nova] Acquired lock "refresh_cache-8c4b7227-1a01-4400-beb8-2df27e17e329" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.661136] env[61728]: DEBUG nova.network.neutron [req-f416ec17-fa59-4327-b94e-a1a408571a7b req-5eac146d-1387-451f-8e77-7dd42391f4b2 service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Refreshing network info cache for port f77e3955-f8e6-43aa-ad48-d8bc4d078422 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 554.951855] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463560, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.603139} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.955248] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2/a0831461-ece1-43ee-92f6-34d7d4e673e2.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 554.955248] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 554.955248] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d29c8840-5c3f-414b-aaa9-94631d7daf55 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.963044] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 554.963044] env[61728]: value = "task-463562" [ 554.963044] env[61728]: _type = "Task" [ 554.963044] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.979639] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463562, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.028836] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-463561, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.067430] env[61728]: DEBUG nova.network.neutron [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating instance_info_cache with network_info: [{"id": "1d776ec6-69e5-4f30-8201-13c69916cca0", "address": "fa:16:3e:6c:2d:56", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.159", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d776ec6-69", "ovs_interfaceid": "1d776ec6-69e5-4f30-8201-13c69916cca0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.490523] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463562, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.537495] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-463561, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.235779} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.537851] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 8c4b7227-1a01-4400-beb8-2df27e17e329/8c4b7227-1a01-4400-beb8-2df27e17e329.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 555.538115] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 555.538381] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b293106e-4561-4d82-917f-c2b92548b4eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.559017] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Waiting for the task: (returnval){ [ 555.559017] env[61728]: value = "task-463563" [ 555.559017] env[61728]: _type = "Task" [ 555.559017] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.568052] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-463563, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.572857] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Releasing lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.592354] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429ea9fa-5643-43df-8448-92bdd55a069e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.630701] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d68628a-9a0b-40c3-9d47-b170caa80ff6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.668128] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4901d12-75bf-46ea-b574-c9f13f99ca11 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.682019] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62194be9-71bf-421e-9069-0f7c3a597f97 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.698573] env[61728]: DEBUG nova.compute.provider_tree [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 555.811248] env[61728]: DEBUG nova.network.neutron [req-f416ec17-fa59-4327-b94e-a1a408571a7b req-5eac146d-1387-451f-8e77-7dd42391f4b2 service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Updated VIF entry in instance network info cache for port f77e3955-f8e6-43aa-ad48-d8bc4d078422. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 555.814513] env[61728]: DEBUG nova.network.neutron [req-f416ec17-fa59-4327-b94e-a1a408571a7b req-5eac146d-1387-451f-8e77-7dd42391f4b2 service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Updating instance_info_cache with network_info: [{"id": "f77e3955-f8e6-43aa-ad48-d8bc4d078422", "address": "fa:16:3e:af:89:5c", "network": {"id": "77ffa432-b9ec-4b9f-b25b-b7d2366d3b37", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-458592337-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e4f01166b049579fe2527d66e4f768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf77e3955-f8", "ovs_interfaceid": "f77e3955-f8e6-43aa-ad48-d8bc4d078422", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.830952] env[61728]: DEBUG nova.compute.manager [req-dbaf48d0-590e-4eaf-8c42-78e30e85af57 req-8b44e040-559c-4f20-9711-9af0b75f06d2 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Received event network-changed-edc90742-f02d-42ad-85cd-b4242f797c33 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 555.830952] env[61728]: DEBUG nova.compute.manager [req-dbaf48d0-590e-4eaf-8c42-78e30e85af57 req-8b44e040-559c-4f20-9711-9af0b75f06d2 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Refreshing instance network info cache due to event network-changed-edc90742-f02d-42ad-85cd-b4242f797c33. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 555.831529] env[61728]: DEBUG oslo_concurrency.lockutils [req-dbaf48d0-590e-4eaf-8c42-78e30e85af57 req-8b44e040-559c-4f20-9711-9af0b75f06d2 service nova] Acquiring lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.831529] env[61728]: DEBUG oslo_concurrency.lockutils [req-dbaf48d0-590e-4eaf-8c42-78e30e85af57 req-8b44e040-559c-4f20-9711-9af0b75f06d2 service nova] Acquired lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.831732] env[61728]: DEBUG nova.network.neutron [req-dbaf48d0-590e-4eaf-8c42-78e30e85af57 req-8b44e040-559c-4f20-9711-9af0b75f06d2 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Refreshing network info cache for port edc90742-f02d-42ad-85cd-b4242f797c33 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 555.979594] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463562, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.571771} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.980104] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 555.980904] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2006d7b6-49f2-42b4-94de-26b8bab826d8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.009441] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Reconfiguring VM instance instance-00000010 to attach disk [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2/a0831461-ece1-43ee-92f6-34d7d4e673e2.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 556.010867] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a5a644a-e600-4a84-9ba4-316c760028ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.031260] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 556.031260] env[61728]: value = "task-463564" [ 556.031260] env[61728]: _type = "Task" [ 556.031260] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.040273] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463564, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.066431] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-463563, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093188} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.066701] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 556.067512] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63cd6d11-3758-4fbb-8451-9d79bc4277de {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.096270] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Reconfiguring VM instance instance-0000000f to attach disk [datastore1] 8c4b7227-1a01-4400-beb8-2df27e17e329/8c4b7227-1a01-4400-beb8-2df27e17e329.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 556.098870] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32b70db2-0791-4a56-85c1-349b8e472310 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.121129] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Waiting for the task: (returnval){ [ 556.121129] env[61728]: value = "task-463565" [ 556.121129] env[61728]: _type = "Task" [ 556.121129] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.130539] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-463565, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.134524] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4bdb3d-6bea-415b-b519-8e60f6003edb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.157412] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2521729-95e2-448c-be9d-1d61472cfb8f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.169195] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating instance '9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c' progress to 83 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 556.224503] env[61728]: ERROR nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [req-09658b82-e09b-402c-b930-0e54ccf3dd80] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-09658b82-e09b-402c-b930-0e54ccf3dd80"}]} [ 556.251161] env[61728]: DEBUG nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 556.273694] env[61728]: DEBUG nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 556.273694] env[61728]: DEBUG nova.compute.provider_tree [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 556.292215] env[61728]: DEBUG nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 556.316927] env[61728]: DEBUG oslo_concurrency.lockutils [req-f416ec17-fa59-4327-b94e-a1a408571a7b req-5eac146d-1387-451f-8e77-7dd42391f4b2 service nova] Releasing lock "refresh_cache-8c4b7227-1a01-4400-beb8-2df27e17e329" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.317248] env[61728]: DEBUG nova.compute.manager [req-f416ec17-fa59-4327-b94e-a1a408571a7b req-5eac146d-1387-451f-8e77-7dd42391f4b2 service nova] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Received event network-vif-deleted-86d2fa33-d261-42a1-b49d-1fd478f8311b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 556.322225] env[61728]: DEBUG nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 556.333427] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquiring lock "e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.333680] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Lock "e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.547783] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463564, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.634541] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-463565, 'name': ReconfigVM_Task, 'duration_secs': 0.321736} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.634828] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Reconfigured VM instance instance-0000000f to attach disk [datastore1] 8c4b7227-1a01-4400-beb8-2df27e17e329/8c4b7227-1a01-4400-beb8-2df27e17e329.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 556.635495] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1711ce72-8e27-4e68-a580-9e2f86235af0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.647013] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Waiting for the task: (returnval){ [ 556.647013] env[61728]: value = "task-463566" [ 556.647013] env[61728]: _type = "Task" [ 556.647013] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.656956] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-463566, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.678679] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 556.682426] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d00c4626-94a0-414d-86cf-bfdb27e90629 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.693690] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 556.693690] env[61728]: value = "task-463567" [ 556.693690] env[61728]: _type = "Task" [ 556.693690] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.709699] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463567, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.873015] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb732f88-415f-48e0-bb29-52630c399a47 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.882285] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1731fd1f-17b3-420d-a5ac-ee31c9417ff3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.923106] env[61728]: DEBUG nova.network.neutron [req-dbaf48d0-590e-4eaf-8c42-78e30e85af57 req-8b44e040-559c-4f20-9711-9af0b75f06d2 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Updated VIF entry in instance network info cache for port edc90742-f02d-42ad-85cd-b4242f797c33. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 556.923271] env[61728]: DEBUG nova.network.neutron [req-dbaf48d0-590e-4eaf-8c42-78e30e85af57 req-8b44e040-559c-4f20-9711-9af0b75f06d2 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Updating instance_info_cache with network_info: [{"id": "edc90742-f02d-42ad-85cd-b4242f797c33", "address": "fa:16:3e:aa:75:a5", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedc90742-f0", "ovs_interfaceid": "edc90742-f02d-42ad-85cd-b4242f797c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.924927] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3629e5c4-5752-4ea5-9b28-2468dd36fcb9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.935562] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0919f7b7-37af-4f84-9046-741c2d3f81ef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.961789] env[61728]: DEBUG nova.compute.provider_tree [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 557.048042] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463564, 'name': ReconfigVM_Task, 'duration_secs': 0.723755} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.048512] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Reconfigured VM instance instance-00000010 to attach disk [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2/a0831461-ece1-43ee-92f6-34d7d4e673e2.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 557.050330] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0cb846ed-fa3e-4054-94f5-a07fe598de68 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.057946] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 557.057946] env[61728]: value = "task-463568" [ 557.057946] env[61728]: _type = "Task" [ 557.057946] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.067869] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463568, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.159903] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-463566, 'name': Rename_Task, 'duration_secs': 0.157608} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.163180] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 557.163601] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a37350b-fd8b-4140-887f-60dd8d0d5e9d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.177500] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Waiting for the task: (returnval){ [ 557.177500] env[61728]: value = "task-463569" [ 557.177500] env[61728]: _type = "Task" [ 557.177500] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.187473] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-463569, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.208244] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463567, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.323051] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "0fb1192e-99f1-4469-b196-60df7eab8185" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.323051] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "0fb1192e-99f1-4469-b196-60df7eab8185" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.429585] env[61728]: DEBUG oslo_concurrency.lockutils [req-dbaf48d0-590e-4eaf-8c42-78e30e85af57 req-8b44e040-559c-4f20-9711-9af0b75f06d2 service nova] Releasing lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.488374] env[61728]: ERROR nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [req-f003c9c3-2dd9-4dc3-b21a-90b8231539ec] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f003c9c3-2dd9-4dc3-b21a-90b8231539ec"}]} [ 557.513746] env[61728]: DEBUG nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 557.530471] env[61728]: DEBUG nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 557.530713] env[61728]: DEBUG nova.compute.provider_tree [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 557.547268] env[61728]: DEBUG nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 557.572016] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463568, 'name': Rename_Task, 'duration_secs': 0.241979} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.575555] env[61728]: DEBUG nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 557.575555] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 557.575719] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-653bcccf-c83b-4d8b-9433-6434c4c40bba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.585901] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 557.585901] env[61728]: value = "task-463570" [ 557.585901] env[61728]: _type = "Task" [ 557.585901] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.596795] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.691705] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-463569, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.708106] env[61728]: DEBUG oslo_vmware.api [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463567, 'name': PowerOnVM_Task, 'duration_secs': 0.531445} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.708458] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 557.708458] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c54eb8-0c62-46b6-a96b-2588a2f948e7 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating instance '9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c' progress to 100 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 558.106137] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463570, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.138065] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776f8d1f-f538-42ed-813f-af940f5c560d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.146191] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b18c6ca-0e09-4130-aaf4-11c7cc23fc92 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.181186] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b965001-2a9b-4dc5-a9d9-3087803bdd65 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.195352] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f7d34a-0261-4e35-95c9-42a483a6d9cf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.200115] env[61728]: DEBUG oslo_vmware.api [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-463569, 'name': PowerOnVM_Task, 'duration_secs': 0.63821} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.200434] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 558.200628] env[61728]: INFO nova.compute.manager [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Took 9.40 seconds to spawn the instance on the hypervisor. [ 558.202324] env[61728]: DEBUG nova.compute.manager [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 558.202324] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013ce48e-6bff-4e6e-9b2f-2adf0696805e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.213738] env[61728]: DEBUG nova.compute.provider_tree [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 558.236228] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Acquiring lock "4fcc957e-5175-4589-b4da-d8451d92c815" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.236464] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lock "4fcc957e-5175-4589-b4da-d8451d92c815" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.602315] env[61728]: DEBUG oslo_vmware.api [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463570, 'name': PowerOnVM_Task, 'duration_secs': 0.765052} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.602711] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 558.603463] env[61728]: INFO nova.compute.manager [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Took 6.98 seconds to spawn the instance on the hypervisor. [ 558.603463] env[61728]: DEBUG nova.compute.manager [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 558.603967] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047538fd-9b8d-46ba-9dcb-44fc72f6fad1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.746886] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6ef93387-559d-4a65-bfd7-7f19bba361d0 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.746886] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6ef93387-559d-4a65-bfd7-7f19bba361d0 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.746886] env[61728]: DEBUG nova.compute.manager [None req-6ef93387-559d-4a65-bfd7-7f19bba361d0 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 558.746886] env[61728]: INFO nova.compute.manager [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Took 37.86 seconds to build instance. [ 558.746886] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13b0009-9a9e-433c-8908-6d735954385e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.751583] env[61728]: ERROR nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [req-e1a64845-65f9-4711-8cb9-474feccb8900] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e1a64845-65f9-4711-8cb9-474feccb8900"}]} [ 558.759197] env[61728]: DEBUG nova.compute.manager [None req-6ef93387-559d-4a65-bfd7-7f19bba361d0 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61728) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 558.760149] env[61728]: DEBUG nova.objects.instance [None req-6ef93387-559d-4a65-bfd7-7f19bba361d0 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lazy-loading 'flavor' on Instance uuid 26220c0c-f535-4315-b3bb-2e5ac4d2286e {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 558.774497] env[61728]: DEBUG nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 558.796319] env[61728]: DEBUG nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 558.796319] env[61728]: DEBUG nova.compute.provider_tree [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 558.816146] env[61728]: DEBUG nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 558.843658] env[61728]: DEBUG nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 558.983047] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Acquiring lock "69d7af16-7708-4df6-acca-093b6c1df1ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.983047] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Lock "69d7af16-7708-4df6-acca-093b6c1df1ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.129182] env[61728]: INFO nova.compute.manager [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Took 37.01 seconds to build instance. [ 559.255732] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdcca18d-7d2e-46a9-8731-db9f22b3a0a6 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Lock "8c4b7227-1a01-4400-beb8-2df27e17e329" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.704s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.264905] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ef93387-559d-4a65-bfd7-7f19bba361d0 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 559.265550] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd0c1d41-49e6-4e1e-aef6-0da86d3f220d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.277846] env[61728]: DEBUG oslo_vmware.api [None req-6ef93387-559d-4a65-bfd7-7f19bba361d0 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 559.277846] env[61728]: value = "task-463571" [ 559.277846] env[61728]: _type = "Task" [ 559.277846] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.291609] env[61728]: DEBUG oslo_vmware.api [None req-6ef93387-559d-4a65-bfd7-7f19bba361d0 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463571, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.440018] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2544c9d-e2c4-42b3-b6af-040b730c6acc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.448592] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42abfc5f-9da1-43cc-88a1-3508a6fd5060 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.481483] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce701a1-4c2b-4c2e-a9b7-645728283c6c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.491480] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c82385f-16f8-4933-bd2c-6e79e691c5e0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.509562] env[61728]: DEBUG nova.compute.provider_tree [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 559.635344] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4463d368-1c45-4cf3-b3f5-5f33ba592c2d tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Lock "a0831461-ece1-43ee-92f6-34d7d4e673e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.694s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.760394] env[61728]: DEBUG nova.compute.manager [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 559.793453] env[61728]: DEBUG oslo_vmware.api [None req-6ef93387-559d-4a65-bfd7-7f19bba361d0 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463571, 'name': PowerOffVM_Task, 'duration_secs': 0.407575} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.793751] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ef93387-559d-4a65-bfd7-7f19bba361d0 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 559.793938] env[61728]: DEBUG nova.compute.manager [None req-6ef93387-559d-4a65-bfd7-7f19bba361d0 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 559.794970] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c0b073c-78c2-4136-a94f-1cf06189c0b1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.938765] env[61728]: DEBUG nova.compute.manager [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 559.939285] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8613d7e-4144-470f-866a-a7ad439d7497 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.048281] env[61728]: DEBUG nova.scheduler.client.report [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 47 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 560.049654] env[61728]: DEBUG nova.compute.provider_tree [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 47 to 48 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 560.050400] env[61728]: DEBUG nova.compute.provider_tree [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 560.140864] env[61728]: DEBUG nova.compute.manager [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 560.243172] env[61728]: INFO nova.compute.manager [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Rebuilding instance [ 560.284193] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.289728] env[61728]: DEBUG nova.compute.manager [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 560.290741] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef2d3fb-ceea-44e1-b4bf-540d87f2a8fb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.314469] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6ef93387-559d-4a65-bfd7-7f19bba361d0 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.570s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.452940] env[61728]: INFO nova.compute.manager [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] instance snapshotting [ 560.456091] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141d1d6f-bc34-4d8a-8310-7d8df0b7b6ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.479028] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243a835b-e2c3-4fc2-ba70-f42420ef615a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.559141] env[61728]: DEBUG oslo_concurrency.lockutils [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 6.605s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.559639] env[61728]: DEBUG nova.compute.manager [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 560.563908] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.381s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.564716] env[61728]: DEBUG nova.objects.instance [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Lazy-loading 'resources' on Instance uuid 08ad3f4a-12a8-48d3-8829-f6744648a95e {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 560.664016] env[61728]: DEBUG oslo_concurrency.lockutils [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.807030] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 560.807030] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b14be15-727c-404b-8f3f-ef2d316eb9a1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.817042] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 560.817042] env[61728]: value = "task-463572" [ 560.817042] env[61728]: _type = "Task" [ 560.817042] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.830945] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463572, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.933692] env[61728]: DEBUG nova.compute.manager [req-cdcf3542-af9c-4c67-b5fd-5378b01eaa86 req-eafe5102-756a-45c8-9639-0b1a1c0acbcb service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Received event network-changed-f77e3955-f8e6-43aa-ad48-d8bc4d078422 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 560.934203] env[61728]: DEBUG nova.compute.manager [req-cdcf3542-af9c-4c67-b5fd-5378b01eaa86 req-eafe5102-756a-45c8-9639-0b1a1c0acbcb service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Refreshing instance network info cache due to event network-changed-f77e3955-f8e6-43aa-ad48-d8bc4d078422. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 560.934330] env[61728]: DEBUG oslo_concurrency.lockutils [req-cdcf3542-af9c-4c67-b5fd-5378b01eaa86 req-eafe5102-756a-45c8-9639-0b1a1c0acbcb service nova] Acquiring lock "refresh_cache-8c4b7227-1a01-4400-beb8-2df27e17e329" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.934482] env[61728]: DEBUG oslo_concurrency.lockutils [req-cdcf3542-af9c-4c67-b5fd-5378b01eaa86 req-eafe5102-756a-45c8-9639-0b1a1c0acbcb service nova] Acquired lock "refresh_cache-8c4b7227-1a01-4400-beb8-2df27e17e329" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.934649] env[61728]: DEBUG nova.network.neutron [req-cdcf3542-af9c-4c67-b5fd-5378b01eaa86 req-eafe5102-756a-45c8-9639-0b1a1c0acbcb service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Refreshing network info cache for port f77e3955-f8e6-43aa-ad48-d8bc4d078422 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 560.991138] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 560.991609] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-92aac15e-8ec9-4e5b-9618-131836d329f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.002065] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 561.002065] env[61728]: value = "task-463573" [ 561.002065] env[61728]: _type = "Task" [ 561.002065] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.011425] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463573, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.073054] env[61728]: DEBUG nova.compute.utils [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 561.075702] env[61728]: DEBUG nova.compute.manager [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Not allocating networking since 'none' was specified. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 561.085975] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.086262] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.086450] env[61728]: DEBUG nova.compute.manager [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Going to confirm migration 1 {{(pid=61728) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 561.328345] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463572, 'name': PowerOffVM_Task, 'duration_secs': 0.322619} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.328618] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 561.328705] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 561.332282] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3227d73b-1ce2-457d-a227-53524e43d286 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.343047] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 561.343047] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-031365be-bb10-4fa1-a8f4-48a5da8c288d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.381925] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 561.382158] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 561.382479] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Deleting the datastore file [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 561.382637] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-09340679-4380-492c-b9f1-cc80f9376e11 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.396508] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 561.396508] env[61728]: value = "task-463575" [ 561.396508] env[61728]: _type = "Task" [ 561.396508] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.410385] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463575, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.517326] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463573, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.576467] env[61728]: DEBUG nova.compute.manager [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 561.671818] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a48612-daf0-479d-9326-05c36da09243 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.676494] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.676635] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.676875] env[61728]: DEBUG nova.network.neutron [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 561.677092] env[61728]: DEBUG nova.objects.instance [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lazy-loading 'info_cache' on Instance uuid 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 561.681836] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9568668f-ecc8-4f52-b08b-d5925fe7bd5e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.722534] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c046bf97-fabd-4c91-958e-05c077d16360 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.732472] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5361979-d89c-4aaa-bd49-0ee9d5ade5ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.751404] env[61728]: DEBUG nova.compute.provider_tree [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 561.854055] env[61728]: DEBUG nova.objects.instance [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lazy-loading 'flavor' on Instance uuid 26220c0c-f535-4315-b3bb-2e5ac4d2286e {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 561.866067] env[61728]: DEBUG nova.network.neutron [req-cdcf3542-af9c-4c67-b5fd-5378b01eaa86 req-eafe5102-756a-45c8-9639-0b1a1c0acbcb service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Updated VIF entry in instance network info cache for port f77e3955-f8e6-43aa-ad48-d8bc4d078422. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 561.866920] env[61728]: DEBUG nova.network.neutron [req-cdcf3542-af9c-4c67-b5fd-5378b01eaa86 req-eafe5102-756a-45c8-9639-0b1a1c0acbcb service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Updating instance_info_cache with network_info: [{"id": "f77e3955-f8e6-43aa-ad48-d8bc4d078422", "address": "fa:16:3e:af:89:5c", "network": {"id": "77ffa432-b9ec-4b9f-b25b-b7d2366d3b37", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-458592337-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e4f01166b049579fe2527d66e4f768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf77e3955-f8", "ovs_interfaceid": "f77e3955-f8e6-43aa-ad48-d8bc4d078422", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.906967] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463575, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.114928} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.907587] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 561.908119] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 561.908119] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 562.015047] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463573, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.289884] env[61728]: DEBUG nova.scheduler.client.report [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 48 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 562.290024] env[61728]: DEBUG nova.compute.provider_tree [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 48 to 49 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 562.290412] env[61728]: DEBUG nova.compute.provider_tree [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 562.362649] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "refresh_cache-26220c0c-f535-4315-b3bb-2e5ac4d2286e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.362649] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquired lock "refresh_cache-26220c0c-f535-4315-b3bb-2e5ac4d2286e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.362649] env[61728]: DEBUG nova.network.neutron [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 562.362649] env[61728]: DEBUG nova.objects.instance [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lazy-loading 'info_cache' on Instance uuid 26220c0c-f535-4315-b3bb-2e5ac4d2286e {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 562.368892] env[61728]: DEBUG oslo_concurrency.lockutils [req-cdcf3542-af9c-4c67-b5fd-5378b01eaa86 req-eafe5102-756a-45c8-9639-0b1a1c0acbcb service nova] Releasing lock "refresh_cache-8c4b7227-1a01-4400-beb8-2df27e17e329" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.514066] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463573, 'name': CreateSnapshot_Task, 'duration_secs': 1.278711} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.514341] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 562.515099] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cedab34-d979-4de5-8902-507705af6fd1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.588851] env[61728]: DEBUG nova.compute.manager [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 562.621144] env[61728]: DEBUG nova.virt.hardware [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 562.621144] env[61728]: DEBUG nova.virt.hardware [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 562.621357] env[61728]: DEBUG nova.virt.hardware [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 562.621425] env[61728]: DEBUG nova.virt.hardware [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 562.621609] env[61728]: DEBUG nova.virt.hardware [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 562.621801] env[61728]: DEBUG nova.virt.hardware [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 562.622050] env[61728]: DEBUG nova.virt.hardware [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 562.622233] env[61728]: DEBUG nova.virt.hardware [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 562.622471] env[61728]: DEBUG nova.virt.hardware [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 562.622609] env[61728]: DEBUG nova.virt.hardware [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 562.622780] env[61728]: DEBUG nova.virt.hardware [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 562.624124] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9675767-fe90-497c-91f1-184d1346ba63 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.634118] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c42030-efcf-4cf9-9a3f-b4a26a3330d0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.649294] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Instance VIF info [] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 562.656893] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Creating folder: Project (0aaf0492790e4267844c48c3f189c4fb). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 562.657410] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ae9dda6a-d2c3-47da-be70-4451a9d11b09 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.671236] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Created folder: Project (0aaf0492790e4267844c48c3f189c4fb) in parent group-v121913. [ 562.671572] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Creating folder: Instances. Parent ref: group-v121966. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 562.671953] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-42e03b13-b5dd-4666-b5e8-a1a8bd8b9747 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.682985] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Created folder: Instances in parent group-v121966. [ 562.683253] env[61728]: DEBUG oslo.service.loopingcall [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 562.683446] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 562.683646] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-60bf65e5-5538-40da-a3bc-7f29194557a1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.706530] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 562.706530] env[61728]: value = "task-463578" [ 562.706530] env[61728]: _type = "Task" [ 562.706530] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.720249] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463578, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.796122] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.231s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.800021] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.974s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.800021] env[61728]: INFO nova.compute.claims [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 562.820835] env[61728]: INFO nova.scheduler.client.report [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Deleted allocations for instance 08ad3f4a-12a8-48d3-8829-f6744648a95e [ 562.866299] env[61728]: DEBUG nova.objects.base [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Object Instance<26220c0c-f535-4315-b3bb-2e5ac4d2286e> lazy-loaded attributes: flavor,info_cache {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 562.945544] env[61728]: DEBUG nova.virt.hardware [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 562.945832] env[61728]: DEBUG nova.virt.hardware [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 562.945988] env[61728]: DEBUG nova.virt.hardware [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 562.946254] env[61728]: DEBUG nova.virt.hardware [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 562.946406] env[61728]: DEBUG nova.virt.hardware [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 562.946559] env[61728]: DEBUG nova.virt.hardware [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 562.946806] env[61728]: DEBUG nova.virt.hardware [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 562.947281] env[61728]: DEBUG nova.virt.hardware [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 562.947281] env[61728]: DEBUG nova.virt.hardware [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 562.947393] env[61728]: DEBUG nova.virt.hardware [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 562.947562] env[61728]: DEBUG nova.virt.hardware [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 562.948482] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836a959a-947a-474d-a5d9-17864aa4ee54 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.958121] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-264e7ece-4867-47b9-bcf8-6f601d659a69 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.972300] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Instance VIF info [] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 562.977997] env[61728]: DEBUG oslo.service.loopingcall [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 562.979019] env[61728]: DEBUG nova.network.neutron [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating instance_info_cache with network_info: [{"id": "1d776ec6-69e5-4f30-8201-13c69916cca0", "address": "fa:16:3e:6c:2d:56", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.159", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d776ec6-69", "ovs_interfaceid": "1d776ec6-69e5-4f30-8201-13c69916cca0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.980694] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 562.980917] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b4900eaa-2552-438c-9bbb-182ab4d6ddfc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.000631] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 563.000631] env[61728]: value = "task-463579" [ 563.000631] env[61728]: _type = "Task" [ 563.000631] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.008613] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463579, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.034690] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 563.035368] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3402a6c0-c46d-4acf-90c5-284cc6b5b15a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.044680] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 563.044680] env[61728]: value = "task-463580" [ 563.044680] env[61728]: _type = "Task" [ 563.044680] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.053390] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463580, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.219122] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463578, 'name': CreateVM_Task, 'duration_secs': 0.270875} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.219364] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 563.220013] env[61728]: DEBUG oslo_concurrency.lockutils [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.222069] env[61728]: DEBUG oslo_concurrency.lockutils [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.222069] env[61728]: DEBUG oslo_concurrency.lockutils [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 563.222069] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c711f37-93ad-417d-b6fa-6f327c1751ec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.226687] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Waiting for the task: (returnval){ [ 563.226687] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aa89c2-f462-824d-108f-e4a3927c9caa" [ 563.226687] env[61728]: _type = "Task" [ 563.226687] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.236783] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aa89c2-f462-824d-108f-e4a3927c9caa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.329980] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d61d87b5-446a-471c-b672-49ced73b1066 tempest-ServerExternalEventsTest-95187472 tempest-ServerExternalEventsTest-95187472-project-member] Lock "08ad3f4a-12a8-48d3-8829-f6744648a95e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.813s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.483138] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Releasing lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.483138] env[61728]: DEBUG nova.objects.instance [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lazy-loading 'migration_context' on Instance uuid 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 563.515941] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463579, 'name': CreateVM_Task, 'duration_secs': 0.444604} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.516222] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 563.518212] env[61728]: DEBUG oslo_concurrency.lockutils [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.555514] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463580, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.622350] env[61728]: DEBUG nova.network.neutron [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Updating instance_info_cache with network_info: [{"id": "89110ed6-c4a4-4f3b-b706-134ce068f320", "address": "fa:16:3e:4c:4f:22", "network": {"id": "4da787ba-7f1e-46fb-9902-d90ec8854a65", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-249168995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f587ea8b8fe432aafd3e0ecba054bd0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48937bd1-23dc-413f-b46b-59bf9e709aa0", "external-id": "nsx-vlan-transportzone-160", "segmentation_id": 160, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89110ed6-c4", "ovs_interfaceid": "89110ed6-c4a4-4f3b-b706-134ce068f320", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.744248] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aa89c2-f462-824d-108f-e4a3927c9caa, 'name': SearchDatastore_Task, 'duration_secs': 0.010899} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.744248] env[61728]: DEBUG oslo_concurrency.lockutils [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.744248] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 563.744248] env[61728]: DEBUG oslo_concurrency.lockutils [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.744402] env[61728]: DEBUG oslo_concurrency.lockutils [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.744571] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 563.744900] env[61728]: DEBUG oslo_concurrency.lockutils [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.745255] env[61728]: DEBUG oslo_concurrency.lockutils [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 563.745527] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e73c53b7-60a5-45b3-9f8f-c0fab7e698b3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.747828] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4705a25e-ba65-4e0f-84e7-1e6af307894d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.757645] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 563.757645] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b8e8df-0458-1386-ce52-64102f802344" [ 563.757645] env[61728]: _type = "Task" [ 563.757645] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.759370] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 563.759370] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 563.766182] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3bfec5a-59ff-40c2-9744-76ce98a4b6d1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.773656] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Waiting for the task: (returnval){ [ 563.773656] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52cfeaf2-b7cf-3f55-e273-e89033d3e948" [ 563.773656] env[61728]: _type = "Task" [ 563.773656] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.777811] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b8e8df-0458-1386-ce52-64102f802344, 'name': SearchDatastore_Task, 'duration_secs': 0.009824} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.781330] env[61728]: DEBUG oslo_concurrency.lockutils [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.781989] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 563.781989] env[61728]: DEBUG oslo_concurrency.lockutils [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.788339] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52cfeaf2-b7cf-3f55-e273-e89033d3e948, 'name': SearchDatastore_Task, 'duration_secs': 0.009512} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.789030] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61987792-d733-407c-8d59-3e96d93e7687 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.795631] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Waiting for the task: (returnval){ [ 563.795631] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a0b863-6019-d432-4c48-1cf6be135367" [ 563.795631] env[61728]: _type = "Task" [ 563.795631] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.805388] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a0b863-6019-d432-4c48-1cf6be135367, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.985593] env[61728]: DEBUG nova.objects.base [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Object Instance<9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c> lazy-loaded attributes: info_cache,migration_context {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 563.986558] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39cadc79-14f3-42ba-a464-37be5beccc6b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.020444] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-146e2040-6915-4ad4-9e51-fc1a0810e496 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.028951] env[61728]: DEBUG oslo_vmware.api [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 564.028951] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529d33d8-3ad4-6882-9dd3-2bf8e8380500" [ 564.028951] env[61728]: _type = "Task" [ 564.028951] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.039732] env[61728]: DEBUG oslo_vmware.api [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529d33d8-3ad4-6882-9dd3-2bf8e8380500, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.057501] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463580, 'name': CloneVM_Task} progress is 95%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.124990] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Releasing lock "refresh_cache-26220c0c-f535-4315-b3bb-2e5ac4d2286e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.306454] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a0b863-6019-d432-4c48-1cf6be135367, 'name': SearchDatastore_Task, 'duration_secs': 0.009358} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.309399] env[61728]: DEBUG oslo_concurrency.lockutils [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.309668] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] be014f77-53ca-42a5-9255-2ba950cca314/be014f77-53ca-42a5-9255-2ba950cca314.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 564.310190] env[61728]: DEBUG oslo_concurrency.lockutils [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.310396] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 564.310632] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-97e14f85-831e-47f8-be07-4758eaba95ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.313588] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-296a9c01-ec2a-42c9-872f-62c15275f853 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.320314] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30aeff3b-511d-43e9-b315-846c9bd59d22 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.324712] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Waiting for the task: (returnval){ [ 564.324712] env[61728]: value = "task-463581" [ 564.324712] env[61728]: _type = "Task" [ 564.324712] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.325145] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 564.325205] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 564.326269] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52643639-2d75-428f-addb-364b61acb163 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.336732] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e087bcfc-35db-4e84-849e-ffffd3380691 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.342147] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 564.342147] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c007b1-8028-9a99-1a68-1d33d9ba6cce" [ 564.342147] env[61728]: _type = "Task" [ 564.342147] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.381388] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': task-463581, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.383737] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720b831b-f9f7-4277-a7c6-e9710cdc5294 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.394009] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c007b1-8028-9a99-1a68-1d33d9ba6cce, 'name': SearchDatastore_Task, 'duration_secs': 0.008889} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.394009] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23c2ef21-9d19-4d6f-8ccf-37d7cefede42 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.398235] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0cef53-19d1-4b44-b143-a952b6abab96 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.405905] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 564.405905] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d39e29-84d7-7c1b-c796-4e6670955e78" [ 564.405905] env[61728]: _type = "Task" [ 564.405905] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.417982] env[61728]: DEBUG nova.compute.provider_tree [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 564.426163] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d39e29-84d7-7c1b-c796-4e6670955e78, 'name': SearchDatastore_Task, 'duration_secs': 0.009827} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.427156] env[61728]: DEBUG oslo_concurrency.lockutils [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.427437] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2/a0831461-ece1-43ee-92f6-34d7d4e673e2.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 564.427733] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3fd56d1b-5fcd-436d-9acb-6838aee9b304 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.443882] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 564.443882] env[61728]: value = "task-463582" [ 564.443882] env[61728]: _type = "Task" [ 564.443882] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.460605] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463582, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.541796] env[61728]: DEBUG oslo_vmware.api [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529d33d8-3ad4-6882-9dd3-2bf8e8380500, 'name': SearchDatastore_Task, 'duration_secs': 0.011488} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.542193] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.559353] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463580, 'name': CloneVM_Task, 'duration_secs': 1.299244} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.559460] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Created linked-clone VM from snapshot [ 564.560287] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d51eec-9a44-4450-a2e2-ca87ac57bc67 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.571963] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Uploading image fc281dc7-d3e7-4282-90bd-99c0ae6ad6f1 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 564.590355] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 564.590666] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a3d02289-cbf0-402b-a162-9b95cbd06266 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.600257] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 564.600257] env[61728]: value = "task-463583" [ 564.600257] env[61728]: _type = "Task" [ 564.600257] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.612250] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463583, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.628903] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 564.629328] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85bf9b7e-8c19-4112-8bba-e18e82591fcc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.642572] env[61728]: DEBUG oslo_vmware.api [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 564.642572] env[61728]: value = "task-463584" [ 564.642572] env[61728]: _type = "Task" [ 564.642572] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.663202] env[61728]: DEBUG oslo_vmware.api [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463584, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.836488] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': task-463581, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460291} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.836780] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] be014f77-53ca-42a5-9255-2ba950cca314/be014f77-53ca-42a5-9255-2ba950cca314.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 564.836953] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 564.837308] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b999361d-8ba8-42fb-87c0-ddfbe0fae6dc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.846666] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Waiting for the task: (returnval){ [ 564.846666] env[61728]: value = "task-463585" [ 564.846666] env[61728]: _type = "Task" [ 564.846666] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.860260] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': task-463585, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.920588] env[61728]: DEBUG nova.scheduler.client.report [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 564.955433] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463582, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.118509] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463583, 'name': Destroy_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.156600] env[61728]: DEBUG oslo_vmware.api [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463584, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.358928] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': task-463585, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102478} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.359286] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 565.360138] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dfc6010-cb5f-45d9-96c0-5da6601a2904 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.383807] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Reconfiguring VM instance instance-00000011 to attach disk [datastore1] be014f77-53ca-42a5-9255-2ba950cca314/be014f77-53ca-42a5-9255-2ba950cca314.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 565.384128] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6cff1f1-e8e8-4525-be01-3c396a569b96 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.406549] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Waiting for the task: (returnval){ [ 565.406549] env[61728]: value = "task-463586" [ 565.406549] env[61728]: _type = "Task" [ 565.406549] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.422432] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': task-463586, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.429710] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.632s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.430282] env[61728]: DEBUG nova.compute.manager [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 565.433117] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.534s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.434594] env[61728]: INFO nova.compute.claims [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 565.456279] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463582, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.707009} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.456707] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2/a0831461-ece1-43ee-92f6-34d7d4e673e2.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 565.457408] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 565.457592] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2503bb0f-424d-41a2-8f21-dc00a0a85e97 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.467355] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 565.467355] env[61728]: value = "task-463587" [ 565.467355] env[61728]: _type = "Task" [ 565.467355] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.479850] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463587, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.613686] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463583, 'name': Destroy_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.658450] env[61728]: DEBUG oslo_vmware.api [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463584, 'name': PowerOnVM_Task, 'duration_secs': 0.825314} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.659092] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 565.659302] env[61728]: DEBUG nova.compute.manager [None req-ac98d67a-1638-4994-b23f-50b0f57e9e7e tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 565.660530] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d4c759-875b-4f83-8fac-c94afa33ddb5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.932297] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': task-463586, 'name': ReconfigVM_Task, 'duration_secs': 0.315372} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.932297] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Reconfigured VM instance instance-00000011 to attach disk [datastore1] be014f77-53ca-42a5-9255-2ba950cca314/be014f77-53ca-42a5-9255-2ba950cca314.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 565.932940] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-53919ca6-df7c-49e0-a61d-b4c0438fea2f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.945771] env[61728]: DEBUG nova.compute.utils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 565.950920] env[61728]: DEBUG nova.compute.manager [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 565.951626] env[61728]: DEBUG nova.network.neutron [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 565.954745] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Waiting for the task: (returnval){ [ 565.954745] env[61728]: value = "task-463588" [ 565.954745] env[61728]: _type = "Task" [ 565.954745] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.966969] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': task-463588, 'name': Rename_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.981601] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463587, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.035564] env[61728]: DEBUG nova.policy [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea6bda719ad24902bc0bca7e97be10e7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab547a56a1e24b3bb051b7f0c70b4640', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 566.114201] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463583, 'name': Destroy_Task, 'duration_secs': 1.417167} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.114507] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Destroyed the VM [ 566.114896] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 566.115914] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1ce25889-bf43-470b-9a14-23a69c595321 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.124918] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 566.124918] env[61728]: value = "task-463589" [ 566.124918] env[61728]: _type = "Task" [ 566.124918] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.135284] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463589, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.457597] env[61728]: DEBUG nova.compute.manager [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 566.471205] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': task-463588, 'name': Rename_Task, 'duration_secs': 0.167175} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.474819] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 566.475102] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f2bf6285-55a9-45a5-98c2-cff07dffa3a8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.488643] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463587, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.730045} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.490221] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 566.490764] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Waiting for the task: (returnval){ [ 566.490764] env[61728]: value = "task-463590" [ 566.490764] env[61728]: _type = "Task" [ 566.490764] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.491861] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22fbdba4-79f7-4866-b054-684c87757ad1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.516494] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': task-463590, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.526316] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Reconfiguring VM instance instance-00000010 to attach disk [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2/a0831461-ece1-43ee-92f6-34d7d4e673e2.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 566.530077] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9cf855e4-d587-4b50-b984-850bde095d2e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.554714] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 566.554714] env[61728]: value = "task-463591" [ 566.554714] env[61728]: _type = "Task" [ 566.554714] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.559619] env[61728]: DEBUG nova.network.neutron [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Successfully created port: dd415c0f-ba4e-4019-9b70-d2788151f446 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 566.570175] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463591, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.636886] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463589, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.011021] env[61728]: DEBUG oslo_vmware.api [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': task-463590, 'name': PowerOnVM_Task, 'duration_secs': 0.457531} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.011448] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 567.013946] env[61728]: INFO nova.compute.manager [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Took 4.42 seconds to spawn the instance on the hypervisor. [ 567.013946] env[61728]: DEBUG nova.compute.manager [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 567.013946] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d54954f-8c0f-4db4-a21e-4cded16faddc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.069590] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463591, 'name': ReconfigVM_Task, 'duration_secs': 0.308851} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.071058] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Reconfigured VM instance instance-00000010 to attach disk [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2/a0831461-ece1-43ee-92f6-34d7d4e673e2.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 567.071373] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7eb1a7e2-5ea5-416c-83b5-cbf31a84ba04 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.080698] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 567.080698] env[61728]: value = "task-463592" [ 567.080698] env[61728]: _type = "Task" [ 567.080698] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.105119] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463592, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.124774] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109a3f7f-d695-43e5-9ea6-0e83111e734c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.139679] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1812cd-b24e-4334-b235-55f17fd763ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.148977] env[61728]: DEBUG oslo_vmware.api [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463589, 'name': RemoveSnapshot_Task, 'duration_secs': 0.942514} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.148977] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 567.182627] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc1b515-581e-4710-8b72-f36f4916ae25 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.194936] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8a62a5-d6be-432b-b10f-a38243b0268e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.212415] env[61728]: DEBUG nova.compute.provider_tree [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 567.473541] env[61728]: DEBUG nova.compute.manager [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 567.506028] env[61728]: DEBUG nova.virt.hardware [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:22:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='409468009',id=32,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-36785540',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 567.506258] env[61728]: DEBUG nova.virt.hardware [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 567.506469] env[61728]: DEBUG nova.virt.hardware [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 567.506641] env[61728]: DEBUG nova.virt.hardware [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 567.506692] env[61728]: DEBUG nova.virt.hardware [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 567.506811] env[61728]: DEBUG nova.virt.hardware [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 567.509032] env[61728]: DEBUG nova.virt.hardware [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 567.509032] env[61728]: DEBUG nova.virt.hardware [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 567.509032] env[61728]: DEBUG nova.virt.hardware [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 567.509032] env[61728]: DEBUG nova.virt.hardware [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 567.509032] env[61728]: DEBUG nova.virt.hardware [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 567.509268] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3408bad-7fe2-457d-a3e0-75a054204128 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.517753] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c057b92-82b1-429c-b561-a3f57663a9ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.546484] env[61728]: INFO nova.compute.manager [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Took 41.20 seconds to build instance. [ 567.593102] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463592, 'name': Rename_Task, 'duration_secs': 0.195768} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.593315] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 567.593564] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-69c44a80-1cb1-4c78-bf5a-fca641b22b27 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.604328] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 567.604328] env[61728]: value = "task-463593" [ 567.604328] env[61728]: _type = "Task" [ 567.604328] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.610484] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463593, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.654625] env[61728]: WARNING nova.compute.manager [None req-624b999b-a3f6-4131-9f1a-c5fa18fbc09d tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Image not found during snapshot: nova.exception.ImageNotFound: Image fc281dc7-d3e7-4282-90bd-99c0ae6ad6f1 could not be found. [ 567.751033] env[61728]: DEBUG nova.scheduler.client.report [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 49 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 567.752141] env[61728]: DEBUG nova.compute.provider_tree [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 49 to 50 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 567.752340] env[61728]: DEBUG nova.compute.provider_tree [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 568.049448] env[61728]: DEBUG oslo_concurrency.lockutils [None req-48bc0f1d-cd63-4271-ad5f-2f5e1948d90e tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Lock "be014f77-53ca-42a5-9255-2ba950cca314" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.453s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.114924] env[61728]: DEBUG oslo_vmware.api [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463593, 'name': PowerOnVM_Task, 'duration_secs': 0.452859} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.115399] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 568.116126] env[61728]: DEBUG nova.compute.manager [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 568.118767] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12d0810-56d7-44ec-a416-56ece0b4014a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.258686] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.825s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.259493] env[61728]: DEBUG nova.compute.manager [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 568.263886] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.306s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.264015] env[61728]: DEBUG nova.objects.instance [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lazy-loading 'resources' on Instance uuid 4247e356-ef10-41b3-bb77-106103134d59 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 568.557137] env[61728]: DEBUG nova.compute.manager [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 568.645690] env[61728]: DEBUG oslo_concurrency.lockutils [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.689365] env[61728]: DEBUG nova.network.neutron [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Successfully updated port: dd415c0f-ba4e-4019-9b70-d2788151f446 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 568.767095] env[61728]: DEBUG nova.compute.utils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 568.768937] env[61728]: DEBUG nova.compute.manager [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 568.768937] env[61728]: DEBUG nova.network.neutron [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 568.831530] env[61728]: DEBUG nova.compute.manager [req-24fd2907-84cd-4ff3-8674-269c3552f807 req-6fcb571f-094c-43e4-bd9f-3ff33765e729 service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Received event network-vif-plugged-dd415c0f-ba4e-4019-9b70-d2788151f446 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 568.831996] env[61728]: DEBUG oslo_concurrency.lockutils [req-24fd2907-84cd-4ff3-8674-269c3552f807 req-6fcb571f-094c-43e4-bd9f-3ff33765e729 service nova] Acquiring lock "db8e373d-e06c-43f0-a79c-2c1aaf65a01a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.832132] env[61728]: DEBUG oslo_concurrency.lockutils [req-24fd2907-84cd-4ff3-8674-269c3552f807 req-6fcb571f-094c-43e4-bd9f-3ff33765e729 service nova] Lock "db8e373d-e06c-43f0-a79c-2c1aaf65a01a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.832367] env[61728]: DEBUG oslo_concurrency.lockutils [req-24fd2907-84cd-4ff3-8674-269c3552f807 req-6fcb571f-094c-43e4-bd9f-3ff33765e729 service nova] Lock "db8e373d-e06c-43f0-a79c-2c1aaf65a01a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.832633] env[61728]: DEBUG nova.compute.manager [req-24fd2907-84cd-4ff3-8674-269c3552f807 req-6fcb571f-094c-43e4-bd9f-3ff33765e729 service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] No waiting events found dispatching network-vif-plugged-dd415c0f-ba4e-4019-9b70-d2788151f446 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 568.833041] env[61728]: WARNING nova.compute.manager [req-24fd2907-84cd-4ff3-8674-269c3552f807 req-6fcb571f-094c-43e4-bd9f-3ff33765e729 service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Received unexpected event network-vif-plugged-dd415c0f-ba4e-4019-9b70-d2788151f446 for instance with vm_state building and task_state spawning. [ 568.860512] env[61728]: DEBUG nova.policy [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2a4fae9faff24030b6ffead1ec9b252a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f6788ff3c6ac4aacad4af5e1ba42fb03', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 569.086998] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.193729] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "refresh_cache-db8e373d-e06c-43f0-a79c-2c1aaf65a01a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.194432] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquired lock "refresh_cache-db8e373d-e06c-43f0-a79c-2c1aaf65a01a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.194432] env[61728]: DEBUG nova.network.neutron [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 569.276675] env[61728]: DEBUG nova.compute.manager [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 569.407620] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11e6b9a-9711-41d0-a217-9ca8791e48e7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.418237] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ed869f9-fcff-40d8-8a42-bd66b2d3034c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.454037] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5479596-088e-4ad8-8c34-f1174b91dd6d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.465491] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bdb5006-c006-4951-96c5-3fc01d513042 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.479667] env[61728]: DEBUG nova.compute.provider_tree [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.491043] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "3c4f7d53-6ba6-4e2c-b324-e651640b304b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.491043] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "3c4f7d53-6ba6-4e2c-b324-e651640b304b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.491043] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "3c4f7d53-6ba6-4e2c-b324-e651640b304b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.491043] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "3c4f7d53-6ba6-4e2c-b324-e651640b304b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.492670] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "3c4f7d53-6ba6-4e2c-b324-e651640b304b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.493439] env[61728]: INFO nova.compute.manager [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Terminating instance [ 569.496165] env[61728]: DEBUG nova.compute.manager [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 569.496345] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 569.497164] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050c5af7-4e4f-4713-97ad-833d551d1d25 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.505307] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 569.506113] env[61728]: DEBUG nova.network.neutron [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Successfully created port: 36a914b9-8961-48e0-b44c-25f49ddbade5 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 569.507836] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a6938748-f5d5-442c-a04d-a88ccbd4148b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.515182] env[61728]: DEBUG oslo_vmware.api [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 569.515182] env[61728]: value = "task-463594" [ 569.515182] env[61728]: _type = "Task" [ 569.515182] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.525565] env[61728]: DEBUG oslo_vmware.api [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463594, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.741663] env[61728]: DEBUG nova.network.neutron [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.439150] env[61728]: DEBUG nova.compute.manager [None req-d8497a92-5589-4b97-9cdb-c167d208a544 tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 570.439815] env[61728]: DEBUG nova.scheduler.client.report [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.444309] env[61728]: DEBUG nova.compute.manager [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 570.453136] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4264df-3716-4108-b47d-fdcbecaca58c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.468998] env[61728]: DEBUG oslo_vmware.api [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463594, 'name': PowerOffVM_Task, 'duration_secs': 0.214269} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.471293] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 570.471474] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 570.472428] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1faae81-be36-46e5-be1e-9554a14e917d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.482943] env[61728]: DEBUG nova.virt.hardware [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 570.483267] env[61728]: DEBUG nova.virt.hardware [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 570.483382] env[61728]: DEBUG nova.virt.hardware [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 570.487021] env[61728]: DEBUG nova.virt.hardware [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 570.487021] env[61728]: DEBUG nova.virt.hardware [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 570.487021] env[61728]: DEBUG nova.virt.hardware [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 570.487021] env[61728]: DEBUG nova.virt.hardware [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 570.487021] env[61728]: DEBUG nova.virt.hardware [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 570.487377] env[61728]: DEBUG nova.virt.hardware [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 570.487377] env[61728]: DEBUG nova.virt.hardware [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 570.487377] env[61728]: DEBUG nova.virt.hardware [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 570.487377] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657c66d5-e9d0-4b4a-9929-e058190dbf11 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.495462] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d004a84f-f4ba-4397-bc23-1fca346df76e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.545396] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 570.545689] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 570.545937] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Deleting the datastore file [datastore1] 3c4f7d53-6ba6-4e2c-b324-e651640b304b {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 570.546311] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-51fcc396-27d6-434c-9494-a9b6a8dcc03e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.552723] env[61728]: DEBUG oslo_vmware.api [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 570.552723] env[61728]: value = "task-463596" [ 570.552723] env[61728]: _type = "Task" [ 570.552723] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.560869] env[61728]: DEBUG oslo_vmware.api [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463596, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.596415] env[61728]: DEBUG nova.network.neutron [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Updating instance_info_cache with network_info: [{"id": "dd415c0f-ba4e-4019-9b70-d2788151f446", "address": "fa:16:3e:95:37:3f", "network": {"id": "113d3399-4dbf-44c8-9fef-d29cbb233c49", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471418633-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab547a56a1e24b3bb051b7f0c70b4640", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd415c0f-ba", "ovs_interfaceid": "dd415c0f-ba4e-4019-9b70-d2788151f446", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.810024] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Acquiring lock "8c53c9c9-cdc1-407b-8827-8409cf137235" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.810359] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Lock "8c53c9c9-cdc1-407b-8827-8409cf137235" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.945798] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.681s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.952156] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.959s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.952156] env[61728]: INFO nova.compute.claims [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 570.954912] env[61728]: INFO nova.compute.manager [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Rebuilding instance [ 570.967407] env[61728]: INFO nova.compute.manager [None req-d8497a92-5589-4b97-9cdb-c167d208a544 tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] instance snapshotting [ 570.968210] env[61728]: DEBUG nova.objects.instance [None req-d8497a92-5589-4b97-9cdb-c167d208a544 tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Lazy-loading 'flavor' on Instance uuid be014f77-53ca-42a5-9255-2ba950cca314 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 570.971699] env[61728]: INFO nova.scheduler.client.report [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleted allocations for instance 4247e356-ef10-41b3-bb77-106103134d59 [ 571.036721] env[61728]: DEBUG nova.compute.manager [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 571.037403] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269dbfa6-7350-4e10-9159-b2af1486c83e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.062735] env[61728]: DEBUG oslo_vmware.api [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463596, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180781} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.062735] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 571.062735] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 571.062735] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 571.062735] env[61728]: INFO nova.compute.manager [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Took 1.57 seconds to destroy the instance on the hypervisor. [ 571.063443] env[61728]: DEBUG oslo.service.loopingcall [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 571.063443] env[61728]: DEBUG nova.compute.manager [-] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 571.063443] env[61728]: DEBUG nova.network.neutron [-] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 571.068866] env[61728]: DEBUG nova.network.neutron [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Successfully updated port: 36a914b9-8961-48e0-b44c-25f49ddbade5 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 571.099742] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Releasing lock "refresh_cache-db8e373d-e06c-43f0-a79c-2c1aaf65a01a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.100387] env[61728]: DEBUG nova.compute.manager [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Instance network_info: |[{"id": "dd415c0f-ba4e-4019-9b70-d2788151f446", "address": "fa:16:3e:95:37:3f", "network": {"id": "113d3399-4dbf-44c8-9fef-d29cbb233c49", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471418633-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab547a56a1e24b3bb051b7f0c70b4640", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd415c0f-ba", "ovs_interfaceid": "dd415c0f-ba4e-4019-9b70-d2788151f446", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 571.100883] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:37:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '459b8c74-0aa6-42b6-996a-42b1c5d7e5c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd415c0f-ba4e-4019-9b70-d2788151f446', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 571.110335] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Creating folder: Project (ab547a56a1e24b3bb051b7f0c70b4640). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 571.112983] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a676ecc8-7503-4773-bc9d-8a655c0abf62 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.116589] env[61728]: DEBUG oslo_concurrency.lockutils [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Acquiring lock "be014f77-53ca-42a5-9255-2ba950cca314" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.116841] env[61728]: DEBUG oslo_concurrency.lockutils [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Lock "be014f77-53ca-42a5-9255-2ba950cca314" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.117166] env[61728]: DEBUG oslo_concurrency.lockutils [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Acquiring lock "be014f77-53ca-42a5-9255-2ba950cca314-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.117716] env[61728]: DEBUG oslo_concurrency.lockutils [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Lock "be014f77-53ca-42a5-9255-2ba950cca314-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.117716] env[61728]: DEBUG oslo_concurrency.lockutils [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Lock "be014f77-53ca-42a5-9255-2ba950cca314-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.119527] env[61728]: INFO nova.compute.manager [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Terminating instance [ 571.121365] env[61728]: DEBUG oslo_concurrency.lockutils [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Acquiring lock "refresh_cache-be014f77-53ca-42a5-9255-2ba950cca314" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.121536] env[61728]: DEBUG oslo_concurrency.lockutils [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Acquired lock "refresh_cache-be014f77-53ca-42a5-9255-2ba950cca314" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.121699] env[61728]: DEBUG nova.network.neutron [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 571.124831] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Created folder: Project (ab547a56a1e24b3bb051b7f0c70b4640) in parent group-v121913. [ 571.125022] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Creating folder: Instances. Parent ref: group-v121971. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 571.125486] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-59632f92-a81a-4804-a2cc-1a6f756cf34a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.136784] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Created folder: Instances in parent group-v121971. [ 571.136784] env[61728]: DEBUG oslo.service.loopingcall [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 571.136784] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 571.137024] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-25191284-1796-4ac3-8da6-c248a08eac2d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.157014] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 571.157014] env[61728]: value = "task-463599" [ 571.157014] env[61728]: _type = "Task" [ 571.157014] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.165471] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463599, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.473541] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c64612e-9b0e-486b-b399-37b3de8a59ce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.494063] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0477de0-7292-475d-bf92-ffc843e670bf tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "4247e356-ef10-41b3-bb77-106103134d59" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.970s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.495524] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fcebc67-138b-41a3-8558-f38cea3e9ad9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.549431] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 571.550602] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a8cb12a-8d9c-4833-af78-93d48bba9399 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.561312] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Waiting for the task: (returnval){ [ 571.561312] env[61728]: value = "task-463600" [ 571.561312] env[61728]: _type = "Task" [ 571.561312] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.571024] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': task-463600, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.574714] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "refresh_cache-c8f1f4ae-ab08-4a03-a3fd-94014509e7e7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.574967] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquired lock "refresh_cache-c8f1f4ae-ab08-4a03-a3fd-94014509e7e7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.575892] env[61728]: DEBUG nova.network.neutron [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 571.609998] env[61728]: DEBUG nova.compute.manager [req-8bbfac30-668f-4b1a-ba86-3d9b612fee0b req-79a9428a-1708-456f-9a04-21c1c95b255b service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Received event network-changed-dd415c0f-ba4e-4019-9b70-d2788151f446 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 571.610279] env[61728]: DEBUG nova.compute.manager [req-8bbfac30-668f-4b1a-ba86-3d9b612fee0b req-79a9428a-1708-456f-9a04-21c1c95b255b service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Refreshing instance network info cache due to event network-changed-dd415c0f-ba4e-4019-9b70-d2788151f446. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 571.610538] env[61728]: DEBUG oslo_concurrency.lockutils [req-8bbfac30-668f-4b1a-ba86-3d9b612fee0b req-79a9428a-1708-456f-9a04-21c1c95b255b service nova] Acquiring lock "refresh_cache-db8e373d-e06c-43f0-a79c-2c1aaf65a01a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.610719] env[61728]: DEBUG oslo_concurrency.lockutils [req-8bbfac30-668f-4b1a-ba86-3d9b612fee0b req-79a9428a-1708-456f-9a04-21c1c95b255b service nova] Acquired lock "refresh_cache-db8e373d-e06c-43f0-a79c-2c1aaf65a01a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.610941] env[61728]: DEBUG nova.network.neutron [req-8bbfac30-668f-4b1a-ba86-3d9b612fee0b req-79a9428a-1708-456f-9a04-21c1c95b255b service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Refreshing network info cache for port dd415c0f-ba4e-4019-9b70-d2788151f446 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 571.623989] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.623989] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.642709] env[61728]: DEBUG nova.network.neutron [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.667626] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463599, 'name': CreateVM_Task, 'duration_secs': 0.495575} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.667803] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 571.668496] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.668655] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.669104] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 571.669313] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e40aa24-8f1c-416c-afe7-e46ce461e688 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.675215] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 571.675215] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fe9ff4-8368-615b-7d1f-652a4df27544" [ 571.675215] env[61728]: _type = "Task" [ 571.675215] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.685844] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fe9ff4-8368-615b-7d1f-652a4df27544, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.762237] env[61728]: DEBUG nova.network.neutron [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.929361] env[61728]: DEBUG nova.network.neutron [-] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.990177] env[61728]: DEBUG nova.compute.manager [req-bc699f40-6fbc-4c3e-9dd9-3455927d908a req-7c68a237-7ccb-457c-a700-7209ee98c46c service nova] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Received event network-vif-deleted-132601da-9151-425d-afd3-3878b0424d7e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 572.007398] env[61728]: DEBUG nova.compute.manager [None req-d8497a92-5589-4b97-9cdb-c167d208a544 tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Instance disappeared during snapshot {{(pid=61728) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 572.083222] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': task-463600, 'name': PowerOffVM_Task, 'duration_secs': 0.143275} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.083222] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 572.083222] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 572.087567] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0973a108-7ab4-4903-85c8-7980848be6b3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.097985] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 572.098793] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2db5fe6e-c25a-48fc-b6bd-4035ce4243d9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.125434] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 572.125434] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 572.125434] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Deleting the datastore file [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 572.125434] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-26bc394f-23e9-47e8-9b83-b6ea0d07cd9d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.127570] env[61728]: DEBUG nova.network.neutron [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 572.137250] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Waiting for the task: (returnval){ [ 572.137250] env[61728]: value = "task-463602" [ 572.137250] env[61728]: _type = "Task" [ 572.137250] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.146964] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': task-463602, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.171483] env[61728]: DEBUG nova.compute.manager [None req-d8497a92-5589-4b97-9cdb-c167d208a544 tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Found 0 images (rotation: 2) {{(pid=61728) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 572.199310] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fe9ff4-8368-615b-7d1f-652a4df27544, 'name': SearchDatastore_Task, 'duration_secs': 0.010785} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.201839] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.202095] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 572.202356] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.202527] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.202758] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 572.203060] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3b37c02-35b1-4e5c-81ae-f5deaf2e0921 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.225392] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 572.225947] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 572.230447] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-676640f9-fe26-4d30-afd4-2e263c05559d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.236651] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 572.236651] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528e8ef8-9d80-b71f-016b-d80dee433c8b" [ 572.236651] env[61728]: _type = "Task" [ 572.236651] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.246615] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528e8ef8-9d80-b71f-016b-d80dee433c8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.266816] env[61728]: DEBUG oslo_concurrency.lockutils [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Releasing lock "refresh_cache-be014f77-53ca-42a5-9255-2ba950cca314" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.267277] env[61728]: DEBUG nova.compute.manager [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 572.267436] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 572.268600] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f8f402-e72b-43ec-b58f-57c0266039a3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.281612] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 572.281895] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88363a23-099d-4074-85f8-5efd882043bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.289563] env[61728]: DEBUG oslo_vmware.api [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Waiting for the task: (returnval){ [ 572.289563] env[61728]: value = "task-463603" [ 572.289563] env[61728]: _type = "Task" [ 572.289563] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.300942] env[61728]: DEBUG oslo_vmware.api [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': task-463603, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.338106] env[61728]: DEBUG nova.network.neutron [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Updating instance_info_cache with network_info: [{"id": "36a914b9-8961-48e0-b44c-25f49ddbade5", "address": "fa:16:3e:78:19:40", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.139", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36a914b9-89", "ovs_interfaceid": "36a914b9-8961-48e0-b44c-25f49ddbade5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.398606] env[61728]: DEBUG nova.network.neutron [req-8bbfac30-668f-4b1a-ba86-3d9b612fee0b req-79a9428a-1708-456f-9a04-21c1c95b255b service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Updated VIF entry in instance network info cache for port dd415c0f-ba4e-4019-9b70-d2788151f446. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 572.399034] env[61728]: DEBUG nova.network.neutron [req-8bbfac30-668f-4b1a-ba86-3d9b612fee0b req-79a9428a-1708-456f-9a04-21c1c95b255b service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Updating instance_info_cache with network_info: [{"id": "dd415c0f-ba4e-4019-9b70-d2788151f446", "address": "fa:16:3e:95:37:3f", "network": {"id": "113d3399-4dbf-44c8-9fef-d29cbb233c49", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471418633-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab547a56a1e24b3bb051b7f0c70b4640", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd415c0f-ba", "ovs_interfaceid": "dd415c0f-ba4e-4019-9b70-d2788151f446", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.435052] env[61728]: INFO nova.compute.manager [-] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Took 1.37 seconds to deallocate network for instance. [ 572.505572] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a1b87c-ba6c-4a81-9db9-1937a7fd5943 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.513141] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f2c491-4d56-4e5c-9991-7a640bb56521 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.543795] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bac386d-cfba-4ed4-81d7-66487de9409c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.556340] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61258f6e-6731-4918-b3b1-addc6840a104 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.569445] env[61728]: DEBUG nova.compute.provider_tree [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.648716] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': task-463602, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196982} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.648782] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 572.648990] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 572.649225] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 572.680159] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "507b5aba-aec9-438b-92c0-5efa17959bc7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.680159] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "507b5aba-aec9-438b-92c0-5efa17959bc7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.746860] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528e8ef8-9d80-b71f-016b-d80dee433c8b, 'name': SearchDatastore_Task, 'duration_secs': 0.015814} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.747649] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4734f8dd-fccb-47bb-a577-863efe242a37 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.752935] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 572.752935] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52657e5d-838d-ec83-100b-415d3e69fc96" [ 572.752935] env[61728]: _type = "Task" [ 572.752935] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.761648] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52657e5d-838d-ec83-100b-415d3e69fc96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.801446] env[61728]: DEBUG oslo_vmware.api [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': task-463603, 'name': PowerOffVM_Task, 'duration_secs': 0.12525} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.801983] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 572.801983] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 572.802187] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b393b9af-dd04-43a0-bac0-6a23487d48f7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.826890] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 572.827145] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 572.827324] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Deleting the datastore file [datastore1] be014f77-53ca-42a5-9255-2ba950cca314 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 572.827575] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-571c0d6f-9edc-4193-b686-fef7a47c786a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.833899] env[61728]: DEBUG oslo_vmware.api [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Waiting for the task: (returnval){ [ 572.833899] env[61728]: value = "task-463605" [ 572.833899] env[61728]: _type = "Task" [ 572.833899] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.842155] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Releasing lock "refresh_cache-c8f1f4ae-ab08-4a03-a3fd-94014509e7e7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.842486] env[61728]: DEBUG nova.compute.manager [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Instance network_info: |[{"id": "36a914b9-8961-48e0-b44c-25f49ddbade5", "address": "fa:16:3e:78:19:40", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.139", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36a914b9-89", "ovs_interfaceid": "36a914b9-8961-48e0-b44c-25f49ddbade5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 572.842756] env[61728]: DEBUG oslo_vmware.api [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': task-463605, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.843128] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:19:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '457c42cd-4ddb-4374-923e-d419b7f6eaff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '36a914b9-8961-48e0-b44c-25f49ddbade5', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 572.850509] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Creating folder: Project (f6788ff3c6ac4aacad4af5e1ba42fb03). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 572.850761] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4da5ed7a-af45-4267-8988-b7933d932149 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.859457] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Created folder: Project (f6788ff3c6ac4aacad4af5e1ba42fb03) in parent group-v121913. [ 572.859677] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Creating folder: Instances. Parent ref: group-v121974. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 572.859864] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d3da33c4-057a-46b3-81ca-22affbe5d343 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.868083] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Created folder: Instances in parent group-v121974. [ 572.868317] env[61728]: DEBUG oslo.service.loopingcall [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 572.868531] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 572.868690] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a3c83ccd-03f6-4fbb-aa66-0e06f00a8152 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.891198] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 572.891198] env[61728]: value = "task-463608" [ 572.891198] env[61728]: _type = "Task" [ 572.891198] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.899076] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463608, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.901734] env[61728]: DEBUG oslo_concurrency.lockutils [req-8bbfac30-668f-4b1a-ba86-3d9b612fee0b req-79a9428a-1708-456f-9a04-21c1c95b255b service nova] Releasing lock "refresh_cache-db8e373d-e06c-43f0-a79c-2c1aaf65a01a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.944029] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.072271] env[61728]: DEBUG nova.scheduler.client.report [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 573.264020] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52657e5d-838d-ec83-100b-415d3e69fc96, 'name': SearchDatastore_Task, 'duration_secs': 0.008997} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.264292] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.264539] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] db8e373d-e06c-43f0-a79c-2c1aaf65a01a/db8e373d-e06c-43f0-a79c-2c1aaf65a01a.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 573.264789] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a2adca78-3d45-4428-ae8b-9728a0b160b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.271391] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 573.271391] env[61728]: value = "task-463609" [ 573.271391] env[61728]: _type = "Task" [ 573.271391] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.278840] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463609, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.343373] env[61728]: DEBUG oslo_vmware.api [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Task: {'id': task-463605, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10816} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.344325] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 573.344325] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 573.344325] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 573.344325] env[61728]: INFO nova.compute.manager [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Took 1.08 seconds to destroy the instance on the hypervisor. [ 573.344517] env[61728]: DEBUG oslo.service.loopingcall [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 573.344566] env[61728]: DEBUG nova.compute.manager [-] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 573.344658] env[61728]: DEBUG nova.network.neutron [-] [instance: be014f77-53ca-42a5-9255-2ba950cca314] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 573.360348] env[61728]: DEBUG nova.network.neutron [-] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.402244] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463608, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.581584] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.632s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.582168] env[61728]: DEBUG nova.compute.manager [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 573.585173] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.795s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.586563] env[61728]: INFO nova.compute.claims [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 573.638062] env[61728]: DEBUG nova.compute.manager [req-1dffce5a-4e14-442c-bcf7-99bc2bcbaceb req-67496fa9-41f8-42ea-905d-64084de55667 service nova] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Received event network-vif-plugged-36a914b9-8961-48e0-b44c-25f49ddbade5 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 573.638062] env[61728]: DEBUG oslo_concurrency.lockutils [req-1dffce5a-4e14-442c-bcf7-99bc2bcbaceb req-67496fa9-41f8-42ea-905d-64084de55667 service nova] Acquiring lock "c8f1f4ae-ab08-4a03-a3fd-94014509e7e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.638302] env[61728]: DEBUG oslo_concurrency.lockutils [req-1dffce5a-4e14-442c-bcf7-99bc2bcbaceb req-67496fa9-41f8-42ea-905d-64084de55667 service nova] Lock "c8f1f4ae-ab08-4a03-a3fd-94014509e7e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.638556] env[61728]: DEBUG oslo_concurrency.lockutils [req-1dffce5a-4e14-442c-bcf7-99bc2bcbaceb req-67496fa9-41f8-42ea-905d-64084de55667 service nova] Lock "c8f1f4ae-ab08-4a03-a3fd-94014509e7e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.638618] env[61728]: DEBUG nova.compute.manager [req-1dffce5a-4e14-442c-bcf7-99bc2bcbaceb req-67496fa9-41f8-42ea-905d-64084de55667 service nova] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] No waiting events found dispatching network-vif-plugged-36a914b9-8961-48e0-b44c-25f49ddbade5 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 573.638739] env[61728]: WARNING nova.compute.manager [req-1dffce5a-4e14-442c-bcf7-99bc2bcbaceb req-67496fa9-41f8-42ea-905d-64084de55667 service nova] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Received unexpected event network-vif-plugged-36a914b9-8961-48e0-b44c-25f49ddbade5 for instance with vm_state building and task_state spawning. [ 573.638901] env[61728]: DEBUG nova.compute.manager [req-1dffce5a-4e14-442c-bcf7-99bc2bcbaceb req-67496fa9-41f8-42ea-905d-64084de55667 service nova] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Received event network-changed-36a914b9-8961-48e0-b44c-25f49ddbade5 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 573.639060] env[61728]: DEBUG nova.compute.manager [req-1dffce5a-4e14-442c-bcf7-99bc2bcbaceb req-67496fa9-41f8-42ea-905d-64084de55667 service nova] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Refreshing instance network info cache due to event network-changed-36a914b9-8961-48e0-b44c-25f49ddbade5. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 573.639250] env[61728]: DEBUG oslo_concurrency.lockutils [req-1dffce5a-4e14-442c-bcf7-99bc2bcbaceb req-67496fa9-41f8-42ea-905d-64084de55667 service nova] Acquiring lock "refresh_cache-c8f1f4ae-ab08-4a03-a3fd-94014509e7e7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.639382] env[61728]: DEBUG oslo_concurrency.lockutils [req-1dffce5a-4e14-442c-bcf7-99bc2bcbaceb req-67496fa9-41f8-42ea-905d-64084de55667 service nova] Acquired lock "refresh_cache-c8f1f4ae-ab08-4a03-a3fd-94014509e7e7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.639557] env[61728]: DEBUG nova.network.neutron [req-1dffce5a-4e14-442c-bcf7-99bc2bcbaceb req-67496fa9-41f8-42ea-905d-64084de55667 service nova] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Refreshing network info cache for port 36a914b9-8961-48e0-b44c-25f49ddbade5 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 573.670962] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "e56a59a6-fc0f-45ca-810a-76b18bae173a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.670962] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "e56a59a6-fc0f-45ca-810a-76b18bae173a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.671211] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "e56a59a6-fc0f-45ca-810a-76b18bae173a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.671585] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "e56a59a6-fc0f-45ca-810a-76b18bae173a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.671676] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "e56a59a6-fc0f-45ca-810a-76b18bae173a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.677803] env[61728]: INFO nova.compute.manager [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Terminating instance [ 573.680049] env[61728]: DEBUG nova.compute.manager [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 573.680206] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 573.681071] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3fe87b-f85a-4303-9725-6c25594c3d44 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.694487] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 573.696899] env[61728]: DEBUG nova.virt.hardware [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 573.697124] env[61728]: DEBUG nova.virt.hardware [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 573.697333] env[61728]: DEBUG nova.virt.hardware [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 573.697565] env[61728]: DEBUG nova.virt.hardware [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 573.697674] env[61728]: DEBUG nova.virt.hardware [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 573.697830] env[61728]: DEBUG nova.virt.hardware [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 573.698051] env[61728]: DEBUG nova.virt.hardware [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 573.698256] env[61728]: DEBUG nova.virt.hardware [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 573.698439] env[61728]: DEBUG nova.virt.hardware [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 573.698603] env[61728]: DEBUG nova.virt.hardware [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 573.698829] env[61728]: DEBUG nova.virt.hardware [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 573.699181] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c579c56b-28d8-4250-bd2b-051dcff81fab {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.702224] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85571724-e8ae-48ee-917f-2776b3632ad9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.712649] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22cd2ecd-ecc9-4a8d-8138-f93bb8f787ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.717117] env[61728]: DEBUG oslo_vmware.api [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 573.717117] env[61728]: value = "task-463610" [ 573.717117] env[61728]: _type = "Task" [ 573.717117] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.731488] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Instance VIF info [] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 573.737647] env[61728]: DEBUG oslo.service.loopingcall [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 573.738537] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 573.738801] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-78662163-c0ae-4f21-909e-6efb53194283 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.755379] env[61728]: DEBUG oslo_vmware.api [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463610, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.762067] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 573.762067] env[61728]: value = "task-463611" [ 573.762067] env[61728]: _type = "Task" [ 573.762067] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.770038] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463611, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.781358] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463609, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502524} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.781630] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] db8e373d-e06c-43f0-a79c-2c1aaf65a01a/db8e373d-e06c-43f0-a79c-2c1aaf65a01a.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 573.781848] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 573.782105] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2082265-6b24-48a5-aa74-f1ced16e40e9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.788179] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 573.788179] env[61728]: value = "task-463612" [ 573.788179] env[61728]: _type = "Task" [ 573.788179] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.797642] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463612, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.863504] env[61728]: DEBUG nova.network.neutron [-] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.903114] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463608, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.094672] env[61728]: DEBUG nova.compute.utils [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 574.098762] env[61728]: DEBUG nova.compute.manager [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Not allocating networking since 'none' was specified. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 574.229867] env[61728]: DEBUG oslo_vmware.api [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463610, 'name': PowerOffVM_Task, 'duration_secs': 0.208885} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.229953] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 574.230095] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 574.230331] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-953e228d-e414-4f8a-b8f2-88d287914f55 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.273018] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463611, 'name': CreateVM_Task, 'duration_secs': 0.418103} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.273198] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 574.273585] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.273741] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.274131] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 574.274374] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17672bf1-ad5e-41f5-b438-3b245f0ea2e0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.278601] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Waiting for the task: (returnval){ [ 574.278601] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d9bf24-f599-a63a-7697-b8fa9f6dd646" [ 574.278601] env[61728]: _type = "Task" [ 574.278601] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.286052] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d9bf24-f599-a63a-7697-b8fa9f6dd646, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.287118] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 574.287318] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 574.287491] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Deleting the datastore file [datastore1] e56a59a6-fc0f-45ca-810a-76b18bae173a {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 574.287717] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1b8c9646-2c0d-49d2-904a-56af8f0f816e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.296805] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463612, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076556} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.297908] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 574.298476] env[61728]: DEBUG oslo_vmware.api [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 574.298476] env[61728]: value = "task-463614" [ 574.298476] env[61728]: _type = "Task" [ 574.298476] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.299338] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c1d485-cc55-408c-aecd-a3e54da0082c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.310293] env[61728]: DEBUG oslo_vmware.api [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463614, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.327642] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Reconfiguring VM instance instance-00000012 to attach disk [datastore1] db8e373d-e06c-43f0-a79c-2c1aaf65a01a/db8e373d-e06c-43f0-a79c-2c1aaf65a01a.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 574.327913] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50d35741-ec46-4f9e-a8ad-0a35a0c909c8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.342505] env[61728]: DEBUG nova.network.neutron [req-1dffce5a-4e14-442c-bcf7-99bc2bcbaceb req-67496fa9-41f8-42ea-905d-64084de55667 service nova] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Updated VIF entry in instance network info cache for port 36a914b9-8961-48e0-b44c-25f49ddbade5. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 574.342943] env[61728]: DEBUG nova.network.neutron [req-1dffce5a-4e14-442c-bcf7-99bc2bcbaceb req-67496fa9-41f8-42ea-905d-64084de55667 service nova] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Updating instance_info_cache with network_info: [{"id": "36a914b9-8961-48e0-b44c-25f49ddbade5", "address": "fa:16:3e:78:19:40", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.139", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36a914b9-89", "ovs_interfaceid": "36a914b9-8961-48e0-b44c-25f49ddbade5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.348308] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 574.348308] env[61728]: value = "task-463615" [ 574.348308] env[61728]: _type = "Task" [ 574.348308] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.356307] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463615, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.366393] env[61728]: INFO nova.compute.manager [-] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Took 1.02 seconds to deallocate network for instance. [ 574.403976] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463608, 'name': CreateVM_Task, 'duration_secs': 1.325035} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.404193] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 574.404879] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.600028] env[61728]: DEBUG nova.compute.manager [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 574.788762] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d9bf24-f599-a63a-7697-b8fa9f6dd646, 'name': SearchDatastore_Task, 'duration_secs': 0.00974} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.791254] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.791495] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 574.791879] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.792055] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.792242] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 574.792706] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.793048] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 574.793276] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29d90b56-760e-4602-800f-4e352ed66184 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.795041] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-774d75ee-735f-4c6b-8003-548cbbe4a4d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.801349] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 574.801349] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d73415-fa52-d9b6-da1d-d55ba9025313" [ 574.801349] env[61728]: _type = "Task" [ 574.801349] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.811727] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 574.811820] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 574.813065] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34e1b926-f5c3-4f20-955f-768d745a5576 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.822558] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d73415-fa52-d9b6-da1d-d55ba9025313, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.822814] env[61728]: DEBUG oslo_vmware.api [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463614, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134728} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.823521] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 574.823742] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 574.823922] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 574.824096] env[61728]: INFO nova.compute.manager [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Took 1.14 seconds to destroy the instance on the hypervisor. [ 574.825062] env[61728]: DEBUG oslo.service.loopingcall [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.825062] env[61728]: DEBUG nova.compute.manager [-] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.825062] env[61728]: DEBUG nova.network.neutron [-] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 574.827251] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Waiting for the task: (returnval){ [ 574.827251] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522c4a45-068b-d27a-14dc-5fb5d3ec3675" [ 574.827251] env[61728]: _type = "Task" [ 574.827251] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.835165] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522c4a45-068b-d27a-14dc-5fb5d3ec3675, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.845221] env[61728]: DEBUG oslo_concurrency.lockutils [req-1dffce5a-4e14-442c-bcf7-99bc2bcbaceb req-67496fa9-41f8-42ea-905d-64084de55667 service nova] Releasing lock "refresh_cache-c8f1f4ae-ab08-4a03-a3fd-94014509e7e7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.857138] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463615, 'name': ReconfigVM_Task, 'duration_secs': 0.249479} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.857404] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Reconfigured VM instance instance-00000012 to attach disk [datastore1] db8e373d-e06c-43f0-a79c-2c1aaf65a01a/db8e373d-e06c-43f0-a79c-2c1aaf65a01a.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 574.860269] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee571a7a-49f4-469f-964c-a684faa02d2a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.866759] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 574.866759] env[61728]: value = "task-463616" [ 574.866759] env[61728]: _type = "Task" [ 574.866759] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.872018] env[61728]: DEBUG oslo_concurrency.lockutils [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.876397] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463616, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.083658] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50235cf-3711-4a58-a8a5-71fa43c83733 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.092615] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde457c9-5904-4246-b6b8-e2dc1630765c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.125968] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a82b755-f05a-436c-b775-181b3daa761e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.133553] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adacb5fe-875a-45c1-a417-c2519ee7d408 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.147291] env[61728]: DEBUG nova.compute.provider_tree [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 575.313699] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d73415-fa52-d9b6-da1d-d55ba9025313, 'name': SearchDatastore_Task, 'duration_secs': 0.044615} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.314047] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.314321] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 575.314482] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.337942] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522c4a45-068b-d27a-14dc-5fb5d3ec3675, 'name': SearchDatastore_Task, 'duration_secs': 0.049576} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.338689] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0072b5cb-1885-47c3-84d7-8d7cd240c9e5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.343721] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Waiting for the task: (returnval){ [ 575.343721] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52208a9d-58de-3da6-57d8-80bab0bcc7e3" [ 575.343721] env[61728]: _type = "Task" [ 575.343721] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.350894] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52208a9d-58de-3da6-57d8-80bab0bcc7e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.374308] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463616, 'name': Rename_Task, 'duration_secs': 0.129112} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.374546] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 575.374763] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c05f0d4-387b-4194-a6da-f160771447ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.380426] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 575.380426] env[61728]: value = "task-463617" [ 575.380426] env[61728]: _type = "Task" [ 575.380426] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.387275] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463617, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.578670] env[61728]: DEBUG nova.network.neutron [-] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.630056] env[61728]: DEBUG nova.compute.manager [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 575.655751] env[61728]: DEBUG nova.virt.hardware [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 575.656046] env[61728]: DEBUG nova.virt.hardware [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 575.656285] env[61728]: DEBUG nova.virt.hardware [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 575.656545] env[61728]: DEBUG nova.virt.hardware [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 575.656734] env[61728]: DEBUG nova.virt.hardware [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 575.656939] env[61728]: DEBUG nova.virt.hardware [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 575.657235] env[61728]: DEBUG nova.virt.hardware [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 575.657439] env[61728]: DEBUG nova.virt.hardware [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 575.657677] env[61728]: DEBUG nova.virt.hardware [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 575.657927] env[61728]: DEBUG nova.virt.hardware [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 575.658216] env[61728]: DEBUG nova.virt.hardware [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 575.659829] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d9f44d-f458-4425-87a5-c00bce271a97 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.668570] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865aa7f6-4c44-406c-bef4-c905670c1b40 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.682655] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Instance VIF info [] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 575.689357] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Creating folder: Project (d32eb64df21c4ccea9882af7dedeb1cc). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 575.690548] env[61728]: ERROR nova.scheduler.client.report [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [req-24285ef8-c01f-4f97-bc91-59b0fbb01925] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-24285ef8-c01f-4f97-bc91-59b0fbb01925"}]} [ 575.690948] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2ff7b6d9-40fe-468b-a7b3-74b197b64358 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.703337] env[61728]: DEBUG nova.compute.manager [req-1ef390f6-fa9a-4620-a907-27e975a2ef72 req-7a081975-abf9-4ba1-a9de-ece6cb38defe service nova] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Received event network-vif-deleted-7af87a49-cf84-45ac-8bb4-e6d6062e4db0 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 575.705793] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Created folder: Project (d32eb64df21c4ccea9882af7dedeb1cc) in parent group-v121913. [ 575.706659] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Creating folder: Instances. Parent ref: group-v121978. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 575.706990] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2356c725-df51-4c94-89da-10f751f011ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.710507] env[61728]: DEBUG nova.scheduler.client.report [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 575.720528] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Created folder: Instances in parent group-v121978. [ 575.720781] env[61728]: DEBUG oslo.service.loopingcall [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 575.720976] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 575.721243] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d55e733-cd90-4e6b-8026-c8e5d1700d44 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.734299] env[61728]: DEBUG nova.scheduler.client.report [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 575.734578] env[61728]: DEBUG nova.compute.provider_tree [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 575.741373] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 575.741373] env[61728]: value = "task-463620" [ 575.741373] env[61728]: _type = "Task" [ 575.741373] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.749769] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463620, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.750784] env[61728]: DEBUG nova.scheduler.client.report [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 575.770774] env[61728]: DEBUG nova.scheduler.client.report [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 575.857873] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52208a9d-58de-3da6-57d8-80bab0bcc7e3, 'name': SearchDatastore_Task, 'duration_secs': 0.047071} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.858154] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.858430] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2/a0831461-ece1-43ee-92f6-34d7d4e673e2.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 575.858741] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.858939] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 575.859188] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af798ee0-5c93-4485-93d6-6ca37d873403 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.861640] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fab30f0d-32f4-4bdc-949e-f709081c42ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.871520] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Waiting for the task: (returnval){ [ 575.871520] env[61728]: value = "task-463621" [ 575.871520] env[61728]: _type = "Task" [ 575.871520] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.875997] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 575.876074] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 575.879324] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c030f00e-8ef5-4736-84ee-631e777e929b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.885545] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': task-463621, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.891294] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 575.891294] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c85b87-5ed9-0bd4-db4e-fc10624cd5d5" [ 575.891294] env[61728]: _type = "Task" [ 575.891294] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.896936] env[61728]: DEBUG oslo_vmware.api [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463617, 'name': PowerOnVM_Task, 'duration_secs': 0.470185} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.902122] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 575.902336] env[61728]: INFO nova.compute.manager [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Took 8.43 seconds to spawn the instance on the hypervisor. [ 575.902518] env[61728]: DEBUG nova.compute.manager [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 575.903258] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47d6504-ad42-4bcf-800c-4d5847d3886b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.912950] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c85b87-5ed9-0bd4-db4e-fc10624cd5d5, 'name': SearchDatastore_Task, 'duration_secs': 0.007731} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.916595] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81b4cc5c-2b1c-4dee-ac69-3bdc4c2cd84f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.922033] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 575.922033] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52705c4b-4cd0-1d0f-9ef6-1c2f3b077c2a" [ 575.922033] env[61728]: _type = "Task" [ 575.922033] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.932215] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52705c4b-4cd0-1d0f-9ef6-1c2f3b077c2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.081181] env[61728]: INFO nova.compute.manager [-] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Took 1.26 seconds to deallocate network for instance. [ 576.252223] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463620, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.300022] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc41b31-edfb-465a-bb6e-0c66e564adf5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.305447] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c5058b-43de-47fe-ac6d-dbdf4bf16a94 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.338224] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa8b0e2-0105-456b-bf23-3b0e6752fd46 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.345794] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8c603a-82dc-4b56-83ad-54cd88b9fc01 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.358871] env[61728]: DEBUG nova.compute.provider_tree [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.381772] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': task-463621, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.441175} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.382036] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2/a0831461-ece1-43ee-92f6-34d7d4e673e2.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 576.382246] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 576.382465] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da695604-dfbf-4345-b9df-2d9de8d70a04 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.388664] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Waiting for the task: (returnval){ [ 576.388664] env[61728]: value = "task-463622" [ 576.388664] env[61728]: _type = "Task" [ 576.388664] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.395769] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': task-463622, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.430136] env[61728]: INFO nova.compute.manager [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Took 45.62 seconds to build instance. [ 576.437228] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52705c4b-4cd0-1d0f-9ef6-1c2f3b077c2a, 'name': SearchDatastore_Task, 'duration_secs': 0.008399} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.437423] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.437679] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] c8f1f4ae-ab08-4a03-a3fd-94014509e7e7/c8f1f4ae-ab08-4a03-a3fd-94014509e7e7.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 576.437953] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-df7b5cfb-b58e-4336-96b6-107361147013 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.444119] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 576.444119] env[61728]: value = "task-463623" [ 576.444119] env[61728]: _type = "Task" [ 576.444119] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.453686] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463623, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.591266] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.756042] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463620, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.862417] env[61728]: DEBUG nova.scheduler.client.report [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 576.899721] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': task-463622, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060569} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.899983] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 576.900772] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bfd3bc9-5299-41e5-8c3f-003ce91e358a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.919813] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Reconfiguring VM instance instance-00000010 to attach disk [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2/a0831461-ece1-43ee-92f6-34d7d4e673e2.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 576.920388] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39cfb3b9-1962-46a7-b150-9006d0f06fa5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.934501] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b81c7ca-252f-456f-9fc6-0f3ec2a1a586 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "db8e373d-e06c-43f0-a79c-2c1aaf65a01a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.225s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.940515] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Waiting for the task: (returnval){ [ 576.940515] env[61728]: value = "task-463624" [ 576.940515] env[61728]: _type = "Task" [ 576.940515] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.951405] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': task-463624, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.957024] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463623, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.421055} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.957282] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] c8f1f4ae-ab08-4a03-a3fd-94014509e7e7/c8f1f4ae-ab08-4a03-a3fd-94014509e7e7.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 576.957482] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 576.957715] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad20699c-d516-4e4c-ae58-a5907eadf301 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.963160] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 576.963160] env[61728]: value = "task-463625" [ 576.963160] env[61728]: _type = "Task" [ 576.963160] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.970651] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463625, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.257035] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463620, 'name': CreateVM_Task, 'duration_secs': 1.30941} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.257035] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 577.257035] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.257035] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.257035] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 577.257498] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35be4c50-6a09-4221-a3c7-85f55a0363d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.260536] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Waiting for the task: (returnval){ [ 577.260536] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52cb5690-4e67-8f04-1eb6-59ead61579c4" [ 577.260536] env[61728]: _type = "Task" [ 577.260536] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.268522] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52cb5690-4e67-8f04-1eb6-59ead61579c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.367958] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.783s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.368533] env[61728]: DEBUG nova.compute.manager [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 577.371132] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 30.442s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.371309] env[61728]: DEBUG nova.objects.instance [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61728) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 577.437557] env[61728]: DEBUG nova.compute.manager [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 577.452286] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': task-463624, 'name': ReconfigVM_Task, 'duration_secs': 0.326042} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.452631] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Reconfigured VM instance instance-00000010 to attach disk [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2/a0831461-ece1-43ee-92f6-34d7d4e673e2.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 577.453268] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-91665443-0347-476b-b2a3-338bc9dc4a7f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.459908] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Waiting for the task: (returnval){ [ 577.459908] env[61728]: value = "task-463626" [ 577.459908] env[61728]: _type = "Task" [ 577.459908] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.470042] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': task-463626, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.474853] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463625, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061987} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.475027] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 577.475786] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067fd0b0-f25d-4135-9f52-c774fc52ea1b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.498453] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Reconfiguring VM instance instance-00000013 to attach disk [datastore1] c8f1f4ae-ab08-4a03-a3fd-94014509e7e7/c8f1f4ae-ab08-4a03-a3fd-94014509e7e7.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 577.498947] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55abeb70-374f-4c71-b43a-fd100149b339 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.517539] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 577.517539] env[61728]: value = "task-463627" [ 577.517539] env[61728]: _type = "Task" [ 577.517539] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.525741] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463627, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.528391] env[61728]: DEBUG nova.compute.manager [req-b4c1b3d1-73ca-45b8-b6df-3491d59f8d57 req-515dedfc-f8b1-4a44-871a-c69f7817d089 service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Received event network-changed-dd415c0f-ba4e-4019-9b70-d2788151f446 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 577.528599] env[61728]: DEBUG nova.compute.manager [req-b4c1b3d1-73ca-45b8-b6df-3491d59f8d57 req-515dedfc-f8b1-4a44-871a-c69f7817d089 service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Refreshing instance network info cache due to event network-changed-dd415c0f-ba4e-4019-9b70-d2788151f446. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 577.528818] env[61728]: DEBUG oslo_concurrency.lockutils [req-b4c1b3d1-73ca-45b8-b6df-3491d59f8d57 req-515dedfc-f8b1-4a44-871a-c69f7817d089 service nova] Acquiring lock "refresh_cache-db8e373d-e06c-43f0-a79c-2c1aaf65a01a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.528961] env[61728]: DEBUG oslo_concurrency.lockutils [req-b4c1b3d1-73ca-45b8-b6df-3491d59f8d57 req-515dedfc-f8b1-4a44-871a-c69f7817d089 service nova] Acquired lock "refresh_cache-db8e373d-e06c-43f0-a79c-2c1aaf65a01a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.529154] env[61728]: DEBUG nova.network.neutron [req-b4c1b3d1-73ca-45b8-b6df-3491d59f8d57 req-515dedfc-f8b1-4a44-871a-c69f7817d089 service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Refreshing network info cache for port dd415c0f-ba4e-4019-9b70-d2788151f446 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 577.771772] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52cb5690-4e67-8f04-1eb6-59ead61579c4, 'name': SearchDatastore_Task, 'duration_secs': 0.008296} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.772104] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.772341] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 577.772625] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.772768] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.772964] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 577.773228] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51862e85-a6d4-43a7-a091-f3f2f1806f85 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.781327] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 577.781537] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 577.782247] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec11b6a4-04e1-490c-ab6b-b6228d9d1e3b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.788428] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Waiting for the task: (returnval){ [ 577.788428] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52171a65-1a32-a929-7328-d34368ba504a" [ 577.788428] env[61728]: _type = "Task" [ 577.788428] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.796233] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52171a65-1a32-a929-7328-d34368ba504a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.879260] env[61728]: DEBUG nova.compute.utils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 577.880627] env[61728]: DEBUG nova.compute.manager [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 577.880793] env[61728]: DEBUG nova.network.neutron [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 577.938832] env[61728]: DEBUG nova.policy [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b447ad91028a45e1a77813535b5b1df8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '770f0f3fe22b4f3db4f59751aad91ba0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 577.960189] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.969049] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': task-463626, 'name': Rename_Task, 'duration_secs': 0.148661} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.969155] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 577.969383] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c321fcf8-713e-4071-9217-0727574de019 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.975089] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Waiting for the task: (returnval){ [ 577.975089] env[61728]: value = "task-463628" [ 577.975089] env[61728]: _type = "Task" [ 577.975089] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.982464] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': task-463628, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.027210] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463627, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.259914] env[61728]: DEBUG nova.network.neutron [req-b4c1b3d1-73ca-45b8-b6df-3491d59f8d57 req-515dedfc-f8b1-4a44-871a-c69f7817d089 service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Updated VIF entry in instance network info cache for port dd415c0f-ba4e-4019-9b70-d2788151f446. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 578.260666] env[61728]: DEBUG nova.network.neutron [req-b4c1b3d1-73ca-45b8-b6df-3491d59f8d57 req-515dedfc-f8b1-4a44-871a-c69f7817d089 service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Updating instance_info_cache with network_info: [{"id": "dd415c0f-ba4e-4019-9b70-d2788151f446", "address": "fa:16:3e:95:37:3f", "network": {"id": "113d3399-4dbf-44c8-9fef-d29cbb233c49", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471418633-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab547a56a1e24b3bb051b7f0c70b4640", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd415c0f-ba", "ovs_interfaceid": "dd415c0f-ba4e-4019-9b70-d2788151f446", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.263931] env[61728]: DEBUG nova.network.neutron [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Successfully created port: 8c01ddea-32d5-4b65-9774-46a15e7c391f {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 578.298729] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52171a65-1a32-a929-7328-d34368ba504a, 'name': SearchDatastore_Task, 'duration_secs': 0.009213} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.299301] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6e3cea9-c10f-400f-ba37-2beefb50005c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.304401] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Waiting for the task: (returnval){ [ 578.304401] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525e3463-ee04-f155-f253-1a697b3da75b" [ 578.304401] env[61728]: _type = "Task" [ 578.304401] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.312334] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525e3463-ee04-f155-f253-1a697b3da75b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.382439] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2340ed12-6a0a-4aed-9f15-6b354cecee32 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.383716] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.756s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.385668] env[61728]: INFO nova.compute.claims [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 578.389286] env[61728]: DEBUG nova.compute.manager [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 578.489018] env[61728]: DEBUG oslo_vmware.api [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Task: {'id': task-463628, 'name': PowerOnVM_Task, 'duration_secs': 0.409193} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.489018] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 578.489018] env[61728]: DEBUG nova.compute.manager [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 578.489018] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8263dc-82ba-4d1e-8500-5cf7374d670c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.531213] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463627, 'name': ReconfigVM_Task, 'duration_secs': 0.764139} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.531213] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Reconfigured VM instance instance-00000013 to attach disk [datastore1] c8f1f4ae-ab08-4a03-a3fd-94014509e7e7/c8f1f4ae-ab08-4a03-a3fd-94014509e7e7.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 578.531469] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3c9a588-48da-41fa-ae58-b1c92e038dd8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.538711] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 578.538711] env[61728]: value = "task-463629" [ 578.538711] env[61728]: _type = "Task" [ 578.538711] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.547744] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463629, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.763629] env[61728]: DEBUG oslo_concurrency.lockutils [req-b4c1b3d1-73ca-45b8-b6df-3491d59f8d57 req-515dedfc-f8b1-4a44-871a-c69f7817d089 service nova] Releasing lock "refresh_cache-db8e373d-e06c-43f0-a79c-2c1aaf65a01a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.815305] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525e3463-ee04-f155-f253-1a697b3da75b, 'name': SearchDatastore_Task, 'duration_secs': 0.009566} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.815445] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.815703] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2c76f45a-b87c-4c3b-9bf9-ba5d67265437/2c76f45a-b87c-4c3b-9bf9-ba5d67265437.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 578.815957] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ada6055f-c754-4426-bdd0-b790716f1f4b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.822655] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Waiting for the task: (returnval){ [ 578.822655] env[61728]: value = "task-463630" [ 578.822655] env[61728]: _type = "Task" [ 578.822655] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.830194] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463630, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.012998] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.054134] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463629, 'name': Rename_Task, 'duration_secs': 0.156058} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.054316] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 579.054452] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-941f3c7a-9168-4eb2-9c32-62605c09e40d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.061274] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 579.061274] env[61728]: value = "task-463631" [ 579.061274] env[61728]: _type = "Task" [ 579.061274] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.073380] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463631, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.332694] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463630, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45705} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.332999] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2c76f45a-b87c-4c3b-9bf9-ba5d67265437/2c76f45a-b87c-4c3b-9bf9-ba5d67265437.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 579.333169] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 579.333403] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14e0a2c1-0ac4-406c-819c-f1607df4400f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.339119] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Waiting for the task: (returnval){ [ 579.339119] env[61728]: value = "task-463632" [ 579.339119] env[61728]: _type = "Task" [ 579.339119] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.346025] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463632, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.400292] env[61728]: DEBUG nova.compute.manager [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 579.429464] env[61728]: DEBUG nova.virt.hardware [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 579.429792] env[61728]: DEBUG nova.virt.hardware [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 579.429962] env[61728]: DEBUG nova.virt.hardware [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 579.430162] env[61728]: DEBUG nova.virt.hardware [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 579.430310] env[61728]: DEBUG nova.virt.hardware [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 579.430457] env[61728]: DEBUG nova.virt.hardware [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 579.430666] env[61728]: DEBUG nova.virt.hardware [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 579.430825] env[61728]: DEBUG nova.virt.hardware [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 579.430987] env[61728]: DEBUG nova.virt.hardware [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 579.431159] env[61728]: DEBUG nova.virt.hardware [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 579.431326] env[61728]: DEBUG nova.virt.hardware [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 579.433552] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd51a87-8054-413b-bc73-6dfd1b597108 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.441486] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d26beb-79aa-4a5c-b905-065fdc4dd4fc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.572306] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463631, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.727099] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquiring lock "a0831461-ece1-43ee-92f6-34d7d4e673e2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.727099] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Lock "a0831461-ece1-43ee-92f6-34d7d4e673e2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.727662] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquiring lock "a0831461-ece1-43ee-92f6-34d7d4e673e2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.727968] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Lock "a0831461-ece1-43ee-92f6-34d7d4e673e2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.728290] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Lock "a0831461-ece1-43ee-92f6-34d7d4e673e2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.730512] env[61728]: INFO nova.compute.manager [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Terminating instance [ 579.732820] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquiring lock "refresh_cache-a0831461-ece1-43ee-92f6-34d7d4e673e2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.733212] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquired lock "refresh_cache-a0831461-ece1-43ee-92f6-34d7d4e673e2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.733568] env[61728]: DEBUG nova.network.neutron [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 579.774233] env[61728]: DEBUG nova.network.neutron [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Successfully updated port: 8c01ddea-32d5-4b65-9774-46a15e7c391f {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 579.848584] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463632, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.23764} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.848857] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 579.849935] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309dc158-071b-4d72-8e3c-8f7725619546 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.857089] env[61728]: DEBUG nova.compute.manager [req-2d02f309-5296-40e2-b624-277d850a169b req-dcd3fc70-4943-44d3-a9d9-ce02d3b43a7b service nova] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Received event network-vif-plugged-8c01ddea-32d5-4b65-9774-46a15e7c391f {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 579.857307] env[61728]: DEBUG oslo_concurrency.lockutils [req-2d02f309-5296-40e2-b624-277d850a169b req-dcd3fc70-4943-44d3-a9d9-ce02d3b43a7b service nova] Acquiring lock "aec920d1-12fb-4d41-9663-b4c2b7e142d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.857536] env[61728]: DEBUG oslo_concurrency.lockutils [req-2d02f309-5296-40e2-b624-277d850a169b req-dcd3fc70-4943-44d3-a9d9-ce02d3b43a7b service nova] Lock "aec920d1-12fb-4d41-9663-b4c2b7e142d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.857724] env[61728]: DEBUG oslo_concurrency.lockutils [req-2d02f309-5296-40e2-b624-277d850a169b req-dcd3fc70-4943-44d3-a9d9-ce02d3b43a7b service nova] Lock "aec920d1-12fb-4d41-9663-b4c2b7e142d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.857854] env[61728]: DEBUG nova.compute.manager [req-2d02f309-5296-40e2-b624-277d850a169b req-dcd3fc70-4943-44d3-a9d9-ce02d3b43a7b service nova] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] No waiting events found dispatching network-vif-plugged-8c01ddea-32d5-4b65-9774-46a15e7c391f {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 579.858031] env[61728]: WARNING nova.compute.manager [req-2d02f309-5296-40e2-b624-277d850a169b req-dcd3fc70-4943-44d3-a9d9-ce02d3b43a7b service nova] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Received unexpected event network-vif-plugged-8c01ddea-32d5-4b65-9774-46a15e7c391f for instance with vm_state building and task_state spawning. [ 579.873905] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Reconfiguring VM instance instance-00000014 to attach disk [datastore1] 2c76f45a-b87c-4c3b-9bf9-ba5d67265437/2c76f45a-b87c-4c3b-9bf9-ba5d67265437.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 579.876769] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0198d36a-45a5-44b6-a215-6ee82c5240b2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.896027] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Waiting for the task: (returnval){ [ 579.896027] env[61728]: value = "task-463633" [ 579.896027] env[61728]: _type = "Task" [ 579.896027] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.902135] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-534f7414-1a03-4f3b-ad0a-64089ece7444 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.911839] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463633, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.915019] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced2f8da-8917-4964-a618-337562787c90 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.946734] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8a9b408-96df-4c3b-adbe-557173c0bb9a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.954367] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9d76fb-825d-4f2d-89e3-619ae831a508 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.968818] env[61728]: DEBUG nova.compute.provider_tree [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 580.071457] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463631, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.253552] env[61728]: DEBUG nova.network.neutron [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.276794] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Acquiring lock "refresh_cache-aec920d1-12fb-4d41-9663-b4c2b7e142d7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.276973] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Acquired lock "refresh_cache-aec920d1-12fb-4d41-9663-b4c2b7e142d7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.277173] env[61728]: DEBUG nova.network.neutron [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 580.319633] env[61728]: DEBUG nova.network.neutron [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.407175] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463633, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.471974] env[61728]: DEBUG nova.scheduler.client.report [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 580.572611] env[61728]: DEBUG oslo_vmware.api [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463631, 'name': PowerOnVM_Task, 'duration_secs': 1.192828} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.572915] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 580.573171] env[61728]: INFO nova.compute.manager [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Took 10.13 seconds to spawn the instance on the hypervisor. [ 580.573411] env[61728]: DEBUG nova.compute.manager [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 580.574436] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e40201f-bdcc-4e83-94e3-98ccb8dc7475 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.810027] env[61728]: DEBUG nova.network.neutron [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.823441] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Releasing lock "refresh_cache-a0831461-ece1-43ee-92f6-34d7d4e673e2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.823530] env[61728]: DEBUG nova.compute.manager [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.823873] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 580.824934] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a05b124-0cbd-49d0-9ec1-36f2f52d7df1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.835833] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 580.836171] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-14d27c41-7bb1-46b9-9d97-52e053e940e1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.844934] env[61728]: DEBUG oslo_vmware.api [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 580.844934] env[61728]: value = "task-463634" [ 580.844934] env[61728]: _type = "Task" [ 580.844934] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.857702] env[61728]: DEBUG oslo_vmware.api [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463634, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.907867] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463633, 'name': ReconfigVM_Task, 'duration_secs': 1.003195} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.909177] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Reconfigured VM instance instance-00000014 to attach disk [datastore1] 2c76f45a-b87c-4c3b-9bf9-ba5d67265437/2c76f45a-b87c-4c3b-9bf9-ba5d67265437.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 580.909177] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-884baf6c-cd24-40c7-bd48-52cea35938f3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.916066] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Waiting for the task: (returnval){ [ 580.916066] env[61728]: value = "task-463635" [ 580.916066] env[61728]: _type = "Task" [ 580.916066] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.932304] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463635, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.961402] env[61728]: DEBUG nova.network.neutron [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Updating instance_info_cache with network_info: [{"id": "8c01ddea-32d5-4b65-9774-46a15e7c391f", "address": "fa:16:3e:d5:81:4c", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.55", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c01ddea-32", "ovs_interfaceid": "8c01ddea-32d5-4b65-9774-46a15e7c391f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.977221] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.593s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.977982] env[61728]: DEBUG nova.compute.manager [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 580.981426] env[61728]: DEBUG oslo_concurrency.lockutils [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.998s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.981738] env[61728]: DEBUG nova.objects.instance [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Lazy-loading 'resources' on Instance uuid b86bf217-d1f8-4d37-aa87-3a2d27d70420 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 581.093626] env[61728]: INFO nova.compute.manager [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Took 49.23 seconds to build instance. [ 581.356111] env[61728]: DEBUG oslo_vmware.api [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463634, 'name': PowerOffVM_Task, 'duration_secs': 0.117435} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.356111] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 581.356111] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 581.356274] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4d27f5e4-0bf3-4709-b762-31defb509112 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.381118] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 581.381331] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 581.381566] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Deleting the datastore file [datastore1] a0831461-ece1-43ee-92f6-34d7d4e673e2 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 581.382021] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf55d32b-f50b-44f5-8617-f8471750ab03 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.388763] env[61728]: DEBUG oslo_vmware.api [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for the task: (returnval){ [ 581.388763] env[61728]: value = "task-463637" [ 581.388763] env[61728]: _type = "Task" [ 581.388763] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.396987] env[61728]: DEBUG oslo_vmware.api [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463637, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.427458] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463635, 'name': Rename_Task, 'duration_secs': 0.155683} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.427458] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 581.427458] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-599ee258-9670-4162-b259-363106998fd7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.432880] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Waiting for the task: (returnval){ [ 581.432880] env[61728]: value = "task-463638" [ 581.432880] env[61728]: _type = "Task" [ 581.432880] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.448261] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463638, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.465169] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Releasing lock "refresh_cache-aec920d1-12fb-4d41-9663-b4c2b7e142d7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.465923] env[61728]: DEBUG nova.compute.manager [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Instance network_info: |[{"id": "8c01ddea-32d5-4b65-9774-46a15e7c391f", "address": "fa:16:3e:d5:81:4c", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.55", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c01ddea-32", "ovs_interfaceid": "8c01ddea-32d5-4b65-9774-46a15e7c391f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 581.466041] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d5:81:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '457c42cd-4ddb-4374-923e-d419b7f6eaff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c01ddea-32d5-4b65-9774-46a15e7c391f', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 581.474038] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Creating folder: Project (770f0f3fe22b4f3db4f59751aad91ba0). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 581.474349] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f2a18d71-2357-497d-95aa-3a878eca6844 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.489274] env[61728]: DEBUG nova.compute.utils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 581.490745] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Created folder: Project (770f0f3fe22b4f3db4f59751aad91ba0) in parent group-v121913. [ 581.491497] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Creating folder: Instances. Parent ref: group-v121981. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 581.491497] env[61728]: DEBUG nova.compute.manager [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 581.491728] env[61728]: DEBUG nova.network.neutron [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 581.493701] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-23076afb-d4e9-45e3-8db7-9ba40c8f27eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.504672] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Created folder: Instances in parent group-v121981. [ 581.504672] env[61728]: DEBUG oslo.service.loopingcall [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 581.504672] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 581.504824] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cf7dea32-7722-4361-bd56-231e31a69994 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.527035] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 581.527035] env[61728]: value = "task-463641" [ 581.527035] env[61728]: _type = "Task" [ 581.527035] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.536639] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463641, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.545378] env[61728]: DEBUG nova.policy [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '489b3e9916b8461a845e656c7ce2bd2f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cad2b09b986d4096ad91489a99b8d27e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 581.598023] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bda05ef-51dc-4f9b-b631-a519526819ae tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "c8f1f4ae-ab08-4a03-a3fd-94014509e7e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.498s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.832970] env[61728]: DEBUG nova.network.neutron [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Successfully created port: d850f13b-cef8-4936-a8a4-f070c7b531e8 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 581.889839] env[61728]: DEBUG nova.compute.manager [req-6d8d3868-7474-4980-b3e1-5e1c0e9258ed req-745075f1-8801-4f6f-a93c-e00487224e91 service nova] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Received event network-changed-8c01ddea-32d5-4b65-9774-46a15e7c391f {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 581.890057] env[61728]: DEBUG nova.compute.manager [req-6d8d3868-7474-4980-b3e1-5e1c0e9258ed req-745075f1-8801-4f6f-a93c-e00487224e91 service nova] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Refreshing instance network info cache due to event network-changed-8c01ddea-32d5-4b65-9774-46a15e7c391f. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 581.890278] env[61728]: DEBUG oslo_concurrency.lockutils [req-6d8d3868-7474-4980-b3e1-5e1c0e9258ed req-745075f1-8801-4f6f-a93c-e00487224e91 service nova] Acquiring lock "refresh_cache-aec920d1-12fb-4d41-9663-b4c2b7e142d7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.890420] env[61728]: DEBUG oslo_concurrency.lockutils [req-6d8d3868-7474-4980-b3e1-5e1c0e9258ed req-745075f1-8801-4f6f-a93c-e00487224e91 service nova] Acquired lock "refresh_cache-aec920d1-12fb-4d41-9663-b4c2b7e142d7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.890668] env[61728]: DEBUG nova.network.neutron [req-6d8d3868-7474-4980-b3e1-5e1c0e9258ed req-745075f1-8801-4f6f-a93c-e00487224e91 service nova] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Refreshing network info cache for port 8c01ddea-32d5-4b65-9774-46a15e7c391f {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 581.905327] env[61728]: DEBUG oslo_vmware.api [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Task: {'id': task-463637, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108676} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.905327] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 581.905327] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 581.905327] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 581.905327] env[61728]: INFO nova.compute.manager [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Took 1.08 seconds to destroy the instance on the hypervisor. [ 581.905692] env[61728]: DEBUG oslo.service.loopingcall [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 581.905692] env[61728]: DEBUG nova.compute.manager [-] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 581.905692] env[61728]: DEBUG nova.network.neutron [-] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 581.924981] env[61728]: DEBUG nova.network.neutron [-] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.943541] env[61728]: DEBUG oslo_vmware.api [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463638, 'name': PowerOnVM_Task, 'duration_secs': 0.434147} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.947039] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 581.947039] env[61728]: INFO nova.compute.manager [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Took 6.32 seconds to spawn the instance on the hypervisor. [ 581.947039] env[61728]: DEBUG nova.compute.manager [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 581.947970] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655e6baa-5ef2-4d67-b1fd-1cc7e1a878a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.997365] env[61728]: DEBUG nova.compute.manager [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 582.042528] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463641, 'name': CreateVM_Task, 'duration_secs': 0.34273} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.042833] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 582.043636] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.043998] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.044428] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 582.044773] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71c8810c-b79e-4a74-983c-ad01d2c99632 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.052025] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Waiting for the task: (returnval){ [ 582.052025] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527e1e57-532d-2a4a-1d36-df770d82341a" [ 582.052025] env[61728]: _type = "Task" [ 582.052025] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.063806] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527e1e57-532d-2a4a-1d36-df770d82341a, 'name': SearchDatastore_Task, 'duration_secs': 0.009224} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.064236] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.064566] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 582.065121] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.065378] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.067867] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 582.067867] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ceeaf7b4-cef1-4999-917e-7c105f61fb53 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.075412] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667fbd34-914e-4c2b-baac-937523c1bbe4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.076385] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 582.076567] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 582.077584] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f4c7a16-a915-42b9-a2ad-316b506c6dd5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.083163] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac72b73-73e3-463a-a8bf-1dba31986ab5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.087393] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Waiting for the task: (returnval){ [ 582.087393] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5214f8af-cfe2-54fe-cbe8-b16b66754a0d" [ 582.087393] env[61728]: _type = "Task" [ 582.087393] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.123702] env[61728]: DEBUG nova.compute.manager [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 582.129157] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb4bfe1-4a2e-4c45-8b38-a4edddacfa8d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.134976] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5214f8af-cfe2-54fe-cbe8-b16b66754a0d, 'name': SearchDatastore_Task, 'duration_secs': 0.009371} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.136296] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64efc389-cdf0-40f0-a385-0a97c1fdddd3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.142646] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8ca825-73eb-46e7-baee-72fd14b4f6e4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.147499] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Waiting for the task: (returnval){ [ 582.147499] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52472851-cf2a-68ce-fe50-b858df90bc30" [ 582.147499] env[61728]: _type = "Task" [ 582.147499] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.161760] env[61728]: DEBUG nova.compute.provider_tree [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 582.168540] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52472851-cf2a-68ce-fe50-b858df90bc30, 'name': SearchDatastore_Task, 'duration_secs': 0.008194} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.168794] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.169654] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] aec920d1-12fb-4d41-9663-b4c2b7e142d7/aec920d1-12fb-4d41-9663-b4c2b7e142d7.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 582.169877] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e82bfcfa-bb9f-489c-8f7a-e23c34467ca2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.180366] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Waiting for the task: (returnval){ [ 582.180366] env[61728]: value = "task-463642" [ 582.180366] env[61728]: _type = "Task" [ 582.180366] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.189316] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463642, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.427858] env[61728]: DEBUG nova.network.neutron [-] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.442638] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "650d2875-68f7-4131-b9b7-9799ee08f1d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.442901] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "650d2875-68f7-4131-b9b7-9799ee08f1d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.465923] env[61728]: INFO nova.compute.manager [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Took 40.50 seconds to build instance. [ 582.655538] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.666016] env[61728]: DEBUG nova.scheduler.client.report [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.690834] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463642, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458614} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.692062] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] aec920d1-12fb-4d41-9663-b4c2b7e142d7/aec920d1-12fb-4d41-9663-b4c2b7e142d7.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 582.692062] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 582.692203] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d14aa351-a9db-444a-8a68-7a490b74e631 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.700238] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Waiting for the task: (returnval){ [ 582.700238] env[61728]: value = "task-463643" [ 582.700238] env[61728]: _type = "Task" [ 582.700238] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.710617] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463643, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.723365] env[61728]: DEBUG nova.compute.manager [None req-4262abd4-cb8d-430b-a998-596ad6d62913 tempest-ServerDiagnosticsV248Test-360654998 tempest-ServerDiagnosticsV248Test-360654998-project-admin] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 582.727260] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884a37bb-a629-42c3-98ee-c790c6d9eaac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.734396] env[61728]: INFO nova.compute.manager [None req-4262abd4-cb8d-430b-a998-596ad6d62913 tempest-ServerDiagnosticsV248Test-360654998 tempest-ServerDiagnosticsV248Test-360654998-project-admin] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Retrieving diagnostics [ 582.735493] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44935d3c-05cf-477f-a084-7f5ba0ca2a9b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.783949] env[61728]: DEBUG nova.network.neutron [req-6d8d3868-7474-4980-b3e1-5e1c0e9258ed req-745075f1-8801-4f6f-a93c-e00487224e91 service nova] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Updated VIF entry in instance network info cache for port 8c01ddea-32d5-4b65-9774-46a15e7c391f. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 582.785057] env[61728]: DEBUG nova.network.neutron [req-6d8d3868-7474-4980-b3e1-5e1c0e9258ed req-745075f1-8801-4f6f-a93c-e00487224e91 service nova] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Updating instance_info_cache with network_info: [{"id": "8c01ddea-32d5-4b65-9774-46a15e7c391f", "address": "fa:16:3e:d5:81:4c", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.55", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c01ddea-32", "ovs_interfaceid": "8c01ddea-32d5-4b65-9774-46a15e7c391f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.932228] env[61728]: INFO nova.compute.manager [-] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Took 1.03 seconds to deallocate network for instance. [ 582.967922] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e181228a-4a0b-41a1-83f6-6e26a3274a1a tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Lock "2c76f45a-b87c-4c3b-9bf9-ba5d67265437" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.083s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.006617] env[61728]: DEBUG nova.compute.manager [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 583.030696] env[61728]: DEBUG nova.virt.hardware [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 583.030963] env[61728]: DEBUG nova.virt.hardware [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 583.031152] env[61728]: DEBUG nova.virt.hardware [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 583.031344] env[61728]: DEBUG nova.virt.hardware [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 583.031511] env[61728]: DEBUG nova.virt.hardware [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 583.031689] env[61728]: DEBUG nova.virt.hardware [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 583.031922] env[61728]: DEBUG nova.virt.hardware [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 583.032117] env[61728]: DEBUG nova.virt.hardware [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 583.032297] env[61728]: DEBUG nova.virt.hardware [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 583.032484] env[61728]: DEBUG nova.virt.hardware [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 583.032698] env[61728]: DEBUG nova.virt.hardware [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 583.033610] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2312479-b41a-4b19-97ef-7b4093fcd7ef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.041489] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf509769-773d-4ea8-863b-dd27d4634fcf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.175299] env[61728]: DEBUG oslo_concurrency.lockutils [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.194s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.178190] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.234s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.179290] env[61728]: INFO nova.compute.claims [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 583.197807] env[61728]: INFO nova.scheduler.client.report [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Deleted allocations for instance b86bf217-d1f8-4d37-aa87-3a2d27d70420 [ 583.212648] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463643, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067651} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.212904] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 583.213685] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8780f30a-360e-40e8-9412-9a9b2ca11fc2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.239854] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Reconfiguring VM instance instance-00000015 to attach disk [datastore1] aec920d1-12fb-4d41-9663-b4c2b7e142d7/aec920d1-12fb-4d41-9663-b4c2b7e142d7.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 583.240316] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57c20106-b8c0-46fe-888e-6aa5f20f156c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.263196] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Waiting for the task: (returnval){ [ 583.263196] env[61728]: value = "task-463644" [ 583.263196] env[61728]: _type = "Task" [ 583.263196] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.272746] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463644, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.286710] env[61728]: DEBUG oslo_concurrency.lockutils [req-6d8d3868-7474-4980-b3e1-5e1c0e9258ed req-745075f1-8801-4f6f-a93c-e00487224e91 service nova] Releasing lock "refresh_cache-aec920d1-12fb-4d41-9663-b4c2b7e142d7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.372798] env[61728]: DEBUG nova.network.neutron [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Successfully updated port: d850f13b-cef8-4936-a8a4-f070c7b531e8 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 583.442020] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.471113] env[61728]: DEBUG nova.compute.manager [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 583.709996] env[61728]: DEBUG oslo_concurrency.lockutils [None req-104af8ca-b653-4342-b596-4e7bc550a104 tempest-ServerDiagnosticsNegativeTest-713105694 tempest-ServerDiagnosticsNegativeTest-713105694-project-member] Lock "b86bf217-d1f8-4d37-aa87-3a2d27d70420" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.009s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.774291] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463644, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.875716] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "refresh_cache-2d389d5e-a53f-424e-90af-5ed05e2f93c7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.875841] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "refresh_cache-2d389d5e-a53f-424e-90af-5ed05e2f93c7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.875977] env[61728]: DEBUG nova.network.neutron [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 583.916367] env[61728]: DEBUG nova.compute.manager [req-a7b13ecb-0d32-4dc6-876c-40f8671f8a51 req-afd8586c-f348-4846-8b26-f2decbbb334b service nova] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Received event network-vif-plugged-d850f13b-cef8-4936-a8a4-f070c7b531e8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 583.916596] env[61728]: DEBUG oslo_concurrency.lockutils [req-a7b13ecb-0d32-4dc6-876c-40f8671f8a51 req-afd8586c-f348-4846-8b26-f2decbbb334b service nova] Acquiring lock "2d389d5e-a53f-424e-90af-5ed05e2f93c7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.917055] env[61728]: DEBUG oslo_concurrency.lockutils [req-a7b13ecb-0d32-4dc6-876c-40f8671f8a51 req-afd8586c-f348-4846-8b26-f2decbbb334b service nova] Lock "2d389d5e-a53f-424e-90af-5ed05e2f93c7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.917264] env[61728]: DEBUG oslo_concurrency.lockutils [req-a7b13ecb-0d32-4dc6-876c-40f8671f8a51 req-afd8586c-f348-4846-8b26-f2decbbb334b service nova] Lock "2d389d5e-a53f-424e-90af-5ed05e2f93c7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.917440] env[61728]: DEBUG nova.compute.manager [req-a7b13ecb-0d32-4dc6-876c-40f8671f8a51 req-afd8586c-f348-4846-8b26-f2decbbb334b service nova] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] No waiting events found dispatching network-vif-plugged-d850f13b-cef8-4936-a8a4-f070c7b531e8 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 583.917656] env[61728]: WARNING nova.compute.manager [req-a7b13ecb-0d32-4dc6-876c-40f8671f8a51 req-afd8586c-f348-4846-8b26-f2decbbb334b service nova] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Received unexpected event network-vif-plugged-d850f13b-cef8-4936-a8a4-f070c7b531e8 for instance with vm_state building and task_state spawning. [ 583.917759] env[61728]: DEBUG nova.compute.manager [req-a7b13ecb-0d32-4dc6-876c-40f8671f8a51 req-afd8586c-f348-4846-8b26-f2decbbb334b service nova] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Received event network-changed-d850f13b-cef8-4936-a8a4-f070c7b531e8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 583.917914] env[61728]: DEBUG nova.compute.manager [req-a7b13ecb-0d32-4dc6-876c-40f8671f8a51 req-afd8586c-f348-4846-8b26-f2decbbb334b service nova] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Refreshing instance network info cache due to event network-changed-d850f13b-cef8-4936-a8a4-f070c7b531e8. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 583.918094] env[61728]: DEBUG oslo_concurrency.lockutils [req-a7b13ecb-0d32-4dc6-876c-40f8671f8a51 req-afd8586c-f348-4846-8b26-f2decbbb334b service nova] Acquiring lock "refresh_cache-2d389d5e-a53f-424e-90af-5ed05e2f93c7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.991649] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.276409] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463644, 'name': ReconfigVM_Task, 'duration_secs': 0.977906} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.276718] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Reconfigured VM instance instance-00000015 to attach disk [datastore1] aec920d1-12fb-4d41-9663-b4c2b7e142d7/aec920d1-12fb-4d41-9663-b4c2b7e142d7.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 584.277375] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fbdc8f92-5c58-406f-83f5-0dd97225cd16 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.283808] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Waiting for the task: (returnval){ [ 584.283808] env[61728]: value = "task-463645" [ 584.283808] env[61728]: _type = "Task" [ 584.283808] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.296024] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463645, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.411390] env[61728]: DEBUG nova.network.neutron [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 584.645206] env[61728]: DEBUG nova.network.neutron [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Updating instance_info_cache with network_info: [{"id": "d850f13b-cef8-4936-a8a4-f070c7b531e8", "address": "fa:16:3e:4b:ba:0e", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd850f13b-ce", "ovs_interfaceid": "d850f13b-cef8-4936-a8a4-f070c7b531e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.695785] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de76ff0c-3346-42bc-8d91-104cb2941efd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.703311] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58eb9c3b-47e5-4d52-8793-425893b9b666 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.735931] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d1aa16-4cab-479b-a2d6-b95575bf4aa1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.743182] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-697841a2-f5d0-4922-8b8c-6f84443049a0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.756330] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 584.756895] env[61728]: DEBUG nova.compute.provider_tree [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.758048] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 584.793528] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463645, 'name': Rename_Task, 'duration_secs': 0.171072} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.793811] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 584.794090] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ec916f4-fe68-40bb-95e5-f381962ef5f3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.800832] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Waiting for the task: (returnval){ [ 584.800832] env[61728]: value = "task-463646" [ 584.800832] env[61728]: _type = "Task" [ 584.800832] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.809414] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463646, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.149333] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "refresh_cache-2d389d5e-a53f-424e-90af-5ed05e2f93c7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.149479] env[61728]: DEBUG nova.compute.manager [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Instance network_info: |[{"id": "d850f13b-cef8-4936-a8a4-f070c7b531e8", "address": "fa:16:3e:4b:ba:0e", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd850f13b-ce", "ovs_interfaceid": "d850f13b-cef8-4936-a8a4-f070c7b531e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 585.149840] env[61728]: DEBUG oslo_concurrency.lockutils [req-a7b13ecb-0d32-4dc6-876c-40f8671f8a51 req-afd8586c-f348-4846-8b26-f2decbbb334b service nova] Acquired lock "refresh_cache-2d389d5e-a53f-424e-90af-5ed05e2f93c7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.149975] env[61728]: DEBUG nova.network.neutron [req-a7b13ecb-0d32-4dc6-876c-40f8671f8a51 req-afd8586c-f348-4846-8b26-f2decbbb334b service nova] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Refreshing network info cache for port d850f13b-cef8-4936-a8a4-f070c7b531e8 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 585.151236] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:ba:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f65996a3-f865-4492-9377-cd14ec8b3aae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd850f13b-cef8-4936-a8a4-f070c7b531e8', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 585.158725] env[61728]: DEBUG oslo.service.loopingcall [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 585.159248] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 585.159473] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d02753ea-7445-48a8-b46a-f68643bd9e9b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.179828] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 585.179828] env[61728]: value = "task-463647" [ 585.179828] env[61728]: _type = "Task" [ 585.179828] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.187592] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463647, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.261198] env[61728]: DEBUG nova.scheduler.client.report [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.268733] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 585.268733] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Starting heal instance info cache {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 585.268865] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Rebuilding the list of instances to heal {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 585.310811] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463646, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.691986] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463647, 'name': CreateVM_Task, 'duration_secs': 0.350762} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.691986] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 585.691986] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.691986] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.692630] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 585.692630] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f286448-cbe4-44cd-b51f-ac9e0f648f83 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.699699] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 585.699699] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dcc78d-fd3c-10db-fb16-16eccad33d54" [ 585.699699] env[61728]: _type = "Task" [ 585.699699] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.710607] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dcc78d-fd3c-10db-fb16-16eccad33d54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.769595] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.592s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.770331] env[61728]: DEBUG nova.compute.manager [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 585.773884] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.935s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.775683] env[61728]: INFO nova.compute.claims [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.783824] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Skipping network cache update for instance because it is Building. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 585.791049] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Skipping network cache update for instance because it is Building. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 585.792448] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Skipping network cache update for instance because it is Building. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 585.811480] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463646, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.858234] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.858234] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquired lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.858234] env[61728]: DEBUG nova.network.neutron [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Forcefully refreshing network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 585.858234] env[61728]: DEBUG nova.objects.instance [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lazy-loading 'info_cache' on Instance uuid 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 585.923423] env[61728]: DEBUG nova.network.neutron [req-a7b13ecb-0d32-4dc6-876c-40f8671f8a51 req-afd8586c-f348-4846-8b26-f2decbbb334b service nova] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Updated VIF entry in instance network info cache for port d850f13b-cef8-4936-a8a4-f070c7b531e8. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 585.923423] env[61728]: DEBUG nova.network.neutron [req-a7b13ecb-0d32-4dc6-876c-40f8671f8a51 req-afd8586c-f348-4846-8b26-f2decbbb334b service nova] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Updating instance_info_cache with network_info: [{"id": "d850f13b-cef8-4936-a8a4-f070c7b531e8", "address": "fa:16:3e:4b:ba:0e", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd850f13b-ce", "ovs_interfaceid": "d850f13b-cef8-4936-a8a4-f070c7b531e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.210741] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dcc78d-fd3c-10db-fb16-16eccad33d54, 'name': SearchDatastore_Task, 'duration_secs': 0.009879} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.211043] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.211282] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 586.211551] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.211743] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.211930] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 586.212195] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a0788d55-d899-4f1b-a9e5-d819257391f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.220630] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 586.220814] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 586.221542] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-178032dd-d5c4-4b34-83e8-aed6144db744 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.226867] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 586.226867] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522278db-0860-3747-bd0d-e25112ab5ad6" [ 586.226867] env[61728]: _type = "Task" [ 586.226867] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.236299] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522278db-0860-3747-bd0d-e25112ab5ad6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.298222] env[61728]: DEBUG nova.compute.utils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 586.299807] env[61728]: DEBUG nova.compute.manager [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 586.300071] env[61728]: DEBUG nova.network.neutron [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 586.313918] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463646, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.342605] env[61728]: DEBUG nova.policy [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3375e38dea6c461b9ba5a5aa765ef68d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '62e6ae50d2a34b7280e409111d327a0d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 586.425302] env[61728]: DEBUG oslo_concurrency.lockutils [req-a7b13ecb-0d32-4dc6-876c-40f8671f8a51 req-afd8586c-f348-4846-8b26-f2decbbb334b service nova] Releasing lock "refresh_cache-2d389d5e-a53f-424e-90af-5ed05e2f93c7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.633234] env[61728]: DEBUG nova.network.neutron [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Successfully created port: 6408b220-22b3-4b0d-961e-9a7a9ae42648 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 586.737991] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522278db-0860-3747-bd0d-e25112ab5ad6, 'name': SearchDatastore_Task, 'duration_secs': 0.008637} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.738797] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4aeb6786-d0dc-4be9-ab42-db4fabbaf96e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.745018] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 586.745018] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a8258f-2492-b816-d5bf-187208e6422c" [ 586.745018] env[61728]: _type = "Task" [ 586.745018] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.753100] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a8258f-2492-b816-d5bf-187208e6422c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.805755] env[61728]: DEBUG nova.compute.manager [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 586.825934] env[61728]: DEBUG oslo_vmware.api [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463646, 'name': PowerOnVM_Task, 'duration_secs': 1.543442} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.825934] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 586.826102] env[61728]: INFO nova.compute.manager [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Took 7.43 seconds to spawn the instance on the hypervisor. [ 586.826445] env[61728]: DEBUG nova.compute.manager [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 586.829020] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c01a893d-81c2-4e42-aecf-836864f7db1b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.258169] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a8258f-2492-b816-d5bf-187208e6422c, 'name': SearchDatastore_Task, 'duration_secs': 0.01012} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.261129] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.261186] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2d389d5e-a53f-424e-90af-5ed05e2f93c7/2d389d5e-a53f-424e-90af-5ed05e2f93c7.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 587.261670] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8c0d927-4e13-4e40-806d-c0ce8a4945a5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.268871] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 587.268871] env[61728]: value = "task-463648" [ 587.268871] env[61728]: _type = "Task" [ 587.268871] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.277552] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463648, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.350945] env[61728]: INFO nova.compute.manager [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Took 42.58 seconds to build instance. [ 587.358775] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f403d13-9371-4882-9402-531c9e4db62f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.367506] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8774fb01-cecf-414e-9d04-0b0c1a54a982 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.399324] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67c4f98-f7f3-4aeb-b538-7aaa528fe446 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.406636] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99a99f5-8617-4ffe-96fe-5bc4fc38a6bf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.421572] env[61728]: DEBUG nova.compute.provider_tree [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.787965] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463648, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494133} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.788651] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2d389d5e-a53f-424e-90af-5ed05e2f93c7/2d389d5e-a53f-424e-90af-5ed05e2f93c7.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 587.788651] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 587.788853] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-804d5aff-2c23-46c8-8c3f-f7a92e4bbbc8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.795078] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 587.795078] env[61728]: value = "task-463649" [ 587.795078] env[61728]: _type = "Task" [ 587.795078] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.804225] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463649, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.826261] env[61728]: DEBUG nova.compute.manager [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 587.855337] env[61728]: DEBUG nova.virt.hardware [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 587.855620] env[61728]: DEBUG nova.virt.hardware [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 587.855844] env[61728]: DEBUG nova.virt.hardware [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 587.856075] env[61728]: DEBUG nova.virt.hardware [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 587.856203] env[61728]: DEBUG nova.virt.hardware [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 587.856412] env[61728]: DEBUG nova.virt.hardware [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 587.856642] env[61728]: DEBUG nova.virt.hardware [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 587.856845] env[61728]: DEBUG nova.virt.hardware [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 587.856978] env[61728]: DEBUG nova.virt.hardware [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 587.857163] env[61728]: DEBUG nova.virt.hardware [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 587.857385] env[61728]: DEBUG nova.virt.hardware [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 587.857873] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3856229-f6ca-4271-8c28-a764ff1afc7f tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Lock "aec920d1-12fb-4d41-9663-b4c2b7e142d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.588s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.858688] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9638c7a-7472-4c51-a5f5-699c71ce521b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.867338] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a21af3-2769-45b3-80ed-1ba91d5e0533 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.926053] env[61728]: DEBUG nova.scheduler.client.report [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 588.030785] env[61728]: DEBUG nova.network.neutron [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating instance_info_cache with network_info: [{"id": "1d776ec6-69e5-4f30-8201-13c69916cca0", "address": "fa:16:3e:6c:2d:56", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.159", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d776ec6-69", "ovs_interfaceid": "1d776ec6-69e5-4f30-8201-13c69916cca0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.133306] env[61728]: DEBUG nova.compute.manager [req-4f92557a-886d-40e1-93fa-e7264645c690 req-4e4492a6-fe23-4740-b891-5ad1cecccc87 service nova] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Received event network-vif-plugged-6408b220-22b3-4b0d-961e-9a7a9ae42648 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 588.133306] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f92557a-886d-40e1-93fa-e7264645c690 req-4e4492a6-fe23-4740-b891-5ad1cecccc87 service nova] Acquiring lock "4e5dc620-68a1-4e10-a8be-702c4999ca10-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.133306] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f92557a-886d-40e1-93fa-e7264645c690 req-4e4492a6-fe23-4740-b891-5ad1cecccc87 service nova] Lock "4e5dc620-68a1-4e10-a8be-702c4999ca10-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.133306] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f92557a-886d-40e1-93fa-e7264645c690 req-4e4492a6-fe23-4740-b891-5ad1cecccc87 service nova] Lock "4e5dc620-68a1-4e10-a8be-702c4999ca10-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.133306] env[61728]: DEBUG nova.compute.manager [req-4f92557a-886d-40e1-93fa-e7264645c690 req-4e4492a6-fe23-4740-b891-5ad1cecccc87 service nova] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] No waiting events found dispatching network-vif-plugged-6408b220-22b3-4b0d-961e-9a7a9ae42648 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 588.133498] env[61728]: WARNING nova.compute.manager [req-4f92557a-886d-40e1-93fa-e7264645c690 req-4e4492a6-fe23-4740-b891-5ad1cecccc87 service nova] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Received unexpected event network-vif-plugged-6408b220-22b3-4b0d-961e-9a7a9ae42648 for instance with vm_state building and task_state spawning. [ 588.185629] env[61728]: DEBUG nova.network.neutron [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Successfully updated port: 6408b220-22b3-4b0d-961e-9a7a9ae42648 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 588.305476] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463649, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06555} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.305741] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 588.306566] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eebcbb29-1fac-4868-b75f-4574982a0d29 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.329351] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Reconfiguring VM instance instance-00000016 to attach disk [datastore1] 2d389d5e-a53f-424e-90af-5ed05e2f93c7/2d389d5e-a53f-424e-90af-5ed05e2f93c7.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 588.330557] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce3e7955-2adf-4daf-82e5-ba01dd197fb3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.350093] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 588.350093] env[61728]: value = "task-463650" [ 588.350093] env[61728]: _type = "Task" [ 588.350093] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.357679] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463650, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.364281] env[61728]: DEBUG nova.compute.manager [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 588.432179] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.658s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.432810] env[61728]: DEBUG nova.compute.manager [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 588.439055] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.070s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.439055] env[61728]: DEBUG nova.objects.instance [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lazy-loading 'resources' on Instance uuid ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 588.534708] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Releasing lock "refresh_cache-9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.534822] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updated the network info_cache for instance {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 588.534945] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 588.535944] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 588.535944] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 588.535944] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 588.535944] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 588.535944] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 588.536214] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61728) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 588.536214] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 588.690212] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Acquiring lock "refresh_cache-4e5dc620-68a1-4e10-a8be-702c4999ca10" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.690643] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Acquired lock "refresh_cache-4e5dc620-68a1-4e10-a8be-702c4999ca10" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.690805] env[61728]: DEBUG nova.network.neutron [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 588.860877] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463650, 'name': ReconfigVM_Task, 'duration_secs': 0.287563} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.861168] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Reconfigured VM instance instance-00000016 to attach disk [datastore1] 2d389d5e-a53f-424e-90af-5ed05e2f93c7/2d389d5e-a53f-424e-90af-5ed05e2f93c7.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 588.861853] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a95c1116-f348-4148-9e19-3e0ad4162abf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.870950] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 588.870950] env[61728]: value = "task-463651" [ 588.870950] env[61728]: _type = "Task" [ 588.870950] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.882510] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463651, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.891177] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.940525] env[61728]: DEBUG nova.compute.utils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.942143] env[61728]: DEBUG nova.compute.manager [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 588.942309] env[61728]: DEBUG nova.network.neutron [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 589.008807] env[61728]: DEBUG nova.policy [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31a2286aad9a40da9b16353541a2780f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49c64edaa37f41d38aba7bd5b1d0b47f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 589.041262] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.269236] env[61728]: DEBUG nova.network.neutron [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.368017] env[61728]: DEBUG nova.network.neutron [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Successfully created port: 6c9a6b61-5248-4996-a390-c18edb3ec021 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 589.385446] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463651, 'name': Rename_Task, 'duration_secs': 0.158206} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.385714] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 589.385992] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe502e8a-9522-4b7a-991c-9be2a2cf8f41 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.392145] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 589.392145] env[61728]: value = "task-463652" [ 589.392145] env[61728]: _type = "Task" [ 589.392145] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.410423] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463652, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.448260] env[61728]: DEBUG nova.compute.manager [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 589.464695] env[61728]: DEBUG nova.network.neutron [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Updating instance_info_cache with network_info: [{"id": "6408b220-22b3-4b0d-961e-9a7a9ae42648", "address": "fa:16:3e:31:7a:9d", "network": {"id": "36cd4339-1796-4b57-85f8-950073dcd063", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1977850529-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62e6ae50d2a34b7280e409111d327a0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6408b220-22", "ovs_interfaceid": "6408b220-22b3-4b0d-961e-9a7a9ae42648", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.556101] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc602a2-cd11-4e4a-ba78-d68d0e7ba263 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.566772] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-683d2b92-4177-4e49-9c30-e6033c96359d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.614953] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7812cf5-3b98-4498-b1f6-29977d09bba9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.623216] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909e103f-40e3-47bd-ab14-92da90839ae3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.638688] env[61728]: DEBUG nova.compute.provider_tree [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.649339] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Acquiring lock "ad304c11-4ff8-4910-92a5-24216d727c4b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.649642] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Lock "ad304c11-4ff8-4910-92a5-24216d727c4b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.905803] env[61728]: DEBUG oslo_vmware.api [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463652, 'name': PowerOnVM_Task, 'duration_secs': 0.495265} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.906237] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Acquiring lock "aec920d1-12fb-4d41-9663-b4c2b7e142d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.906490] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Lock "aec920d1-12fb-4d41-9663-b4c2b7e142d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.906726] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Acquiring lock "aec920d1-12fb-4d41-9663-b4c2b7e142d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.907055] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Lock "aec920d1-12fb-4d41-9663-b4c2b7e142d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.907261] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Lock "aec920d1-12fb-4d41-9663-b4c2b7e142d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.909093] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 589.909337] env[61728]: INFO nova.compute.manager [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Took 6.90 seconds to spawn the instance on the hypervisor. [ 589.909626] env[61728]: DEBUG nova.compute.manager [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 589.910150] env[61728]: INFO nova.compute.manager [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Terminating instance [ 589.912239] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73641363-1af4-4271-91f8-768694c13938 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.915990] env[61728]: DEBUG nova.compute.manager [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.916185] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 589.917258] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15426705-6225-4c07-82ab-88a3a1e37bb7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.928544] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 589.928951] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dfda7b13-9435-4b25-9cbc-af386f77579e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.938389] env[61728]: DEBUG oslo_vmware.api [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Waiting for the task: (returnval){ [ 589.938389] env[61728]: value = "task-463653" [ 589.938389] env[61728]: _type = "Task" [ 589.938389] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.942855] env[61728]: DEBUG oslo_vmware.api [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463653, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.969870] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Releasing lock "refresh_cache-4e5dc620-68a1-4e10-a8be-702c4999ca10" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.970202] env[61728]: DEBUG nova.compute.manager [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Instance network_info: |[{"id": "6408b220-22b3-4b0d-961e-9a7a9ae42648", "address": "fa:16:3e:31:7a:9d", "network": {"id": "36cd4339-1796-4b57-85f8-950073dcd063", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1977850529-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62e6ae50d2a34b7280e409111d327a0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6408b220-22", "ovs_interfaceid": "6408b220-22b3-4b0d-961e-9a7a9ae42648", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 589.972267] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:7a:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50cf0a70-948d-4611-af05-94c1483064ed', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6408b220-22b3-4b0d-961e-9a7a9ae42648', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 589.980098] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Creating folder: Project (62e6ae50d2a34b7280e409111d327a0d). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 589.980765] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0e1c6da9-e873-48a8-a364-23cf48ec8ed5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.990645] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Created folder: Project (62e6ae50d2a34b7280e409111d327a0d) in parent group-v121913. [ 589.990824] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Creating folder: Instances. Parent ref: group-v121985. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 589.991062] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f017cd1b-e224-4f6f-926f-4ee2b85b0c29 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.000149] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Created folder: Instances in parent group-v121985. [ 590.000392] env[61728]: DEBUG oslo.service.loopingcall [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 590.000577] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 590.000779] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd5e7532-0911-4277-a878-4ce7d149f631 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.020371] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 590.020371] env[61728]: value = "task-463656" [ 590.020371] env[61728]: _type = "Task" [ 590.020371] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.028693] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463656, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.142286] env[61728]: DEBUG nova.scheduler.client.report [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.187063] env[61728]: DEBUG nova.compute.manager [req-1c07a207-ff12-4938-a609-b1eb0bc8d62e req-995c494e-30a0-4f4d-95af-11593bfeed19 service nova] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Received event network-changed-6408b220-22b3-4b0d-961e-9a7a9ae42648 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 590.187063] env[61728]: DEBUG nova.compute.manager [req-1c07a207-ff12-4938-a609-b1eb0bc8d62e req-995c494e-30a0-4f4d-95af-11593bfeed19 service nova] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Refreshing instance network info cache due to event network-changed-6408b220-22b3-4b0d-961e-9a7a9ae42648. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 590.187063] env[61728]: DEBUG oslo_concurrency.lockutils [req-1c07a207-ff12-4938-a609-b1eb0bc8d62e req-995c494e-30a0-4f4d-95af-11593bfeed19 service nova] Acquiring lock "refresh_cache-4e5dc620-68a1-4e10-a8be-702c4999ca10" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.187063] env[61728]: DEBUG oslo_concurrency.lockutils [req-1c07a207-ff12-4938-a609-b1eb0bc8d62e req-995c494e-30a0-4f4d-95af-11593bfeed19 service nova] Acquired lock "refresh_cache-4e5dc620-68a1-4e10-a8be-702c4999ca10" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.187584] env[61728]: DEBUG nova.network.neutron [req-1c07a207-ff12-4938-a609-b1eb0bc8d62e req-995c494e-30a0-4f4d-95af-11593bfeed19 service nova] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Refreshing network info cache for port 6408b220-22b3-4b0d-961e-9a7a9ae42648 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 590.441028] env[61728]: INFO nova.compute.manager [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Took 41.83 seconds to build instance. [ 590.446156] env[61728]: DEBUG oslo_vmware.api [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463653, 'name': PowerOffVM_Task, 'duration_secs': 0.217713} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.446727] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 590.446890] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 590.447170] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a90c84d8-822f-4529-856b-05b32fbd36ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.460643] env[61728]: DEBUG nova.compute.manager [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 590.488800] env[61728]: DEBUG nova.virt.hardware [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 590.489519] env[61728]: DEBUG nova.virt.hardware [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 590.489519] env[61728]: DEBUG nova.virt.hardware [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 590.489638] env[61728]: DEBUG nova.virt.hardware [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 590.489877] env[61728]: DEBUG nova.virt.hardware [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 590.490111] env[61728]: DEBUG nova.virt.hardware [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 590.490399] env[61728]: DEBUG nova.virt.hardware [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 590.490634] env[61728]: DEBUG nova.virt.hardware [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 590.490888] env[61728]: DEBUG nova.virt.hardware [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 590.491142] env[61728]: DEBUG nova.virt.hardware [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 590.491384] env[61728]: DEBUG nova.virt.hardware [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 590.492851] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3555c250-844b-48fc-ade7-88d747ece48b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.501049] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ac9877-08cd-447a-b9ef-2958bf701bd7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.506442] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 590.506711] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 590.507311] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Deleting the datastore file [datastore1] aec920d1-12fb-4d41-9663-b4c2b7e142d7 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 590.507311] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c05eb6cb-83e7-46f4-abc9-c9931df9bfd4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.521483] env[61728]: DEBUG oslo_vmware.api [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Waiting for the task: (returnval){ [ 590.521483] env[61728]: value = "task-463658" [ 590.521483] env[61728]: _type = "Task" [ 590.521483] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.532617] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463656, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.535665] env[61728]: DEBUG oslo_vmware.api [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463658, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.647944] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.210s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.651229] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.367s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.653031] env[61728]: INFO nova.compute.claims [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 590.674376] env[61728]: INFO nova.scheduler.client.report [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleted allocations for instance ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618 [ 590.890713] env[61728]: DEBUG nova.network.neutron [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Successfully updated port: 6c9a6b61-5248-4996-a390-c18edb3ec021 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 590.945747] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e2f8c2d7-cca7-456d-95d2-cdb54670abd7 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "2d389d5e-a53f-424e-90af-5ed05e2f93c7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.178s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.037842] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463656, 'name': CreateVM_Task, 'duration_secs': 0.622778} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.041784] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 591.042130] env[61728]: DEBUG oslo_vmware.api [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Task: {'id': task-463658, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20603} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.043034] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.043119] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.043390] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 591.043646] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 591.043823] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 591.043999] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 591.044217] env[61728]: INFO nova.compute.manager [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Took 1.13 seconds to destroy the instance on the hypervisor. [ 591.044408] env[61728]: DEBUG oslo.service.loopingcall [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 591.044627] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-daa47a32-ee44-41d1-9eb0-eacde7be6a18 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.050856] env[61728]: DEBUG nova.compute.manager [-] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 591.050974] env[61728]: DEBUG nova.network.neutron [-] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 591.057730] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Waiting for the task: (returnval){ [ 591.057730] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a6b2a0-88c3-67c0-b2dc-dc9c6fe17cab" [ 591.057730] env[61728]: _type = "Task" [ 591.057730] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.064217] env[61728]: DEBUG nova.network.neutron [req-1c07a207-ff12-4938-a609-b1eb0bc8d62e req-995c494e-30a0-4f4d-95af-11593bfeed19 service nova] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Updated VIF entry in instance network info cache for port 6408b220-22b3-4b0d-961e-9a7a9ae42648. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 591.064609] env[61728]: DEBUG nova.network.neutron [req-1c07a207-ff12-4938-a609-b1eb0bc8d62e req-995c494e-30a0-4f4d-95af-11593bfeed19 service nova] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Updating instance_info_cache with network_info: [{"id": "6408b220-22b3-4b0d-961e-9a7a9ae42648", "address": "fa:16:3e:31:7a:9d", "network": {"id": "36cd4339-1796-4b57-85f8-950073dcd063", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1977850529-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62e6ae50d2a34b7280e409111d327a0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6408b220-22", "ovs_interfaceid": "6408b220-22b3-4b0d-961e-9a7a9ae42648", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.069322] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a6b2a0-88c3-67c0-b2dc-dc9c6fe17cab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.182949] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fac9e2ca-a78f-401b-93c9-560b7bb49248 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.593s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.391046] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.391046] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.391949] env[61728]: DEBUG nova.network.neutron [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 591.453101] env[61728]: DEBUG nova.compute.manager [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.570166] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a6b2a0-88c3-67c0-b2dc-dc9c6fe17cab, 'name': SearchDatastore_Task, 'duration_secs': 0.026688} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.570166] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.570166] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 591.570286] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.570357] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.570579] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 591.571152] env[61728]: DEBUG oslo_concurrency.lockutils [req-1c07a207-ff12-4938-a609-b1eb0bc8d62e req-995c494e-30a0-4f4d-95af-11593bfeed19 service nova] Releasing lock "refresh_cache-4e5dc620-68a1-4e10-a8be-702c4999ca10" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.571529] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e8229573-f180-4187-8471-1cd5ff8409b6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.581087] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 591.581302] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 591.583454] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c246857-b5e0-4314-b540-de82c42be30f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.590302] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "b833f691-e76e-4a2e-94a4-7594fadc3ba8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.590630] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "b833f691-e76e-4a2e-94a4-7594fadc3ba8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.594266] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Waiting for the task: (returnval){ [ 591.594266] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b80237-1866-baae-6b68-6af973630fb1" [ 591.594266] env[61728]: _type = "Task" [ 591.594266] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.603419] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b80237-1866-baae-6b68-6af973630fb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.831492] env[61728]: DEBUG nova.network.neutron [-] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.931262] env[61728]: DEBUG nova.network.neutron [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.977740] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.073600] env[61728]: DEBUG nova.network.neutron [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance_info_cache with network_info: [{"id": "6c9a6b61-5248-4996-a390-c18edb3ec021", "address": "fa:16:3e:43:9a:d5", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9a6b61-52", "ovs_interfaceid": "6c9a6b61-5248-4996-a390-c18edb3ec021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.114745] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b80237-1866-baae-6b68-6af973630fb1, 'name': SearchDatastore_Task, 'duration_secs': 0.026466} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.119908] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae2e0107-e968-481a-9af9-15c550d04e5f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.128158] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Waiting for the task: (returnval){ [ 592.128158] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523a20ad-51a3-9484-fd76-77442d3050e5" [ 592.128158] env[61728]: _type = "Task" [ 592.128158] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.140296] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523a20ad-51a3-9484-fd76-77442d3050e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.227695] env[61728]: DEBUG nova.compute.manager [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Received event network-vif-plugged-6c9a6b61-5248-4996-a390-c18edb3ec021 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 592.227899] env[61728]: DEBUG oslo_concurrency.lockutils [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] Acquiring lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.228166] env[61728]: DEBUG oslo_concurrency.lockutils [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] Lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.228295] env[61728]: DEBUG oslo_concurrency.lockutils [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] Lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.228537] env[61728]: DEBUG nova.compute.manager [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] No waiting events found dispatching network-vif-plugged-6c9a6b61-5248-4996-a390-c18edb3ec021 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 592.228668] env[61728]: WARNING nova.compute.manager [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Received unexpected event network-vif-plugged-6c9a6b61-5248-4996-a390-c18edb3ec021 for instance with vm_state building and task_state spawning. [ 592.228735] env[61728]: DEBUG nova.compute.manager [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Received event network-changed-6c9a6b61-5248-4996-a390-c18edb3ec021 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 592.228859] env[61728]: DEBUG nova.compute.manager [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Refreshing instance network info cache due to event network-changed-6c9a6b61-5248-4996-a390-c18edb3ec021. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 592.229032] env[61728]: DEBUG oslo_concurrency.lockutils [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] Acquiring lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.254070] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01b6aac-d6ed-4a8b-b08a-f43e35626f77 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.261869] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d114455-8556-48cc-ae13-513bdcd7b713 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.294347] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e14fe9c6-5be1-4adb-9e1b-6b1611a039c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.302055] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3977165-1127-4f10-8b32-aaf5471fb98f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.316209] env[61728]: DEBUG nova.compute.provider_tree [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.337981] env[61728]: INFO nova.compute.manager [-] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Took 1.29 seconds to deallocate network for instance. [ 592.577327] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.577704] env[61728]: DEBUG nova.compute.manager [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Instance network_info: |[{"id": "6c9a6b61-5248-4996-a390-c18edb3ec021", "address": "fa:16:3e:43:9a:d5", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9a6b61-52", "ovs_interfaceid": "6c9a6b61-5248-4996-a390-c18edb3ec021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 592.578060] env[61728]: DEBUG oslo_concurrency.lockutils [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] Acquired lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.578237] env[61728]: DEBUG nova.network.neutron [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Refreshing network info cache for port 6c9a6b61-5248-4996-a390-c18edb3ec021 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 592.579928] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:9a:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e51ebca-e0f8-4b77-b155-4ff928eef130', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6c9a6b61-5248-4996-a390-c18edb3ec021', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 592.589091] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Creating folder: Project (49c64edaa37f41d38aba7bd5b1d0b47f). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 592.590719] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c3736120-b3ad-4c92-b64c-c1e656905b25 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.601847] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Created folder: Project (49c64edaa37f41d38aba7bd5b1d0b47f) in parent group-v121913. [ 592.602106] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Creating folder: Instances. Parent ref: group-v121988. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 592.602397] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-55f9e6ff-cac4-455e-a394-9549703c3842 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.611413] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Created folder: Instances in parent group-v121988. [ 592.611713] env[61728]: DEBUG oslo.service.loopingcall [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 592.611922] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 592.612198] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f394dfa8-4c27-4cc2-9d1b-5c0c79391d7d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.639099] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523a20ad-51a3-9484-fd76-77442d3050e5, 'name': SearchDatastore_Task, 'duration_secs': 0.010347} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.641568] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.641568] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 4e5dc620-68a1-4e10-a8be-702c4999ca10/4e5dc620-68a1-4e10-a8be-702c4999ca10.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 592.641568] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 592.641568] env[61728]: value = "task-463661" [ 592.641568] env[61728]: _type = "Task" [ 592.641568] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.641568] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b221047d-0465-4b71-86af-f1bc32dd27c8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.652722] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463661, 'name': CreateVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.654201] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Waiting for the task: (returnval){ [ 592.654201] env[61728]: value = "task-463662" [ 592.654201] env[61728]: _type = "Task" [ 592.654201] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.662665] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-463662, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.819095] env[61728]: DEBUG nova.scheduler.client.report [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.845879] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.154680] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463661, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.166919] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-463662, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.310506] env[61728]: DEBUG nova.network.neutron [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updated VIF entry in instance network info cache for port 6c9a6b61-5248-4996-a390-c18edb3ec021. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 593.311175] env[61728]: DEBUG nova.network.neutron [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance_info_cache with network_info: [{"id": "6c9a6b61-5248-4996-a390-c18edb3ec021", "address": "fa:16:3e:43:9a:d5", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9a6b61-52", "ovs_interfaceid": "6c9a6b61-5248-4996-a390-c18edb3ec021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.316391] env[61728]: DEBUG nova.compute.manager [None req-a114bcba-9d9a-489c-877f-4b73ce63e7af tempest-ServerDiagnosticsV248Test-360654998 tempest-ServerDiagnosticsV248Test-360654998-project-admin] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 593.317518] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41936452-89de-460f-90d3-d0ec88d98077 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.324053] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.673s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.324529] env[61728]: DEBUG nova.compute.manager [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 593.328940] env[61728]: DEBUG oslo_concurrency.lockutils [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.665s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.330440] env[61728]: INFO nova.compute.claims [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.333743] env[61728]: INFO nova.compute.manager [None req-a114bcba-9d9a-489c-877f-4b73ce63e7af tempest-ServerDiagnosticsV248Test-360654998 tempest-ServerDiagnosticsV248Test-360654998-project-admin] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Retrieving diagnostics [ 593.334749] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b45958-66a6-4368-84d9-a1495aa246a5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.654750] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463661, 'name': CreateVM_Task, 'duration_secs': 0.639829} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.657662] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 593.658457] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.658635] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.658952] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 593.659532] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b803450e-7937-4cbb-845a-e2c3615019aa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.665249] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-463662, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.560131} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.665848] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 4e5dc620-68a1-4e10-a8be-702c4999ca10/4e5dc620-68a1-4e10-a8be-702c4999ca10.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 593.666067] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 593.666286] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd94d13b-0528-48ed-8610-1366045346e0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.668553] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 593.668553] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524d54cf-b512-2e26-9ec0-c4e7fbc48e7e" [ 593.668553] env[61728]: _type = "Task" [ 593.668553] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.677061] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524d54cf-b512-2e26-9ec0-c4e7fbc48e7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.678154] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Waiting for the task: (returnval){ [ 593.678154] env[61728]: value = "task-463663" [ 593.678154] env[61728]: _type = "Task" [ 593.678154] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.685802] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-463663, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.813830] env[61728]: DEBUG oslo_concurrency.lockutils [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] Releasing lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.814122] env[61728]: DEBUG nova.compute.manager [req-ea5e3ecd-075f-441b-9f02-c3458c907143 req-76f33df0-47f1-4e17-aef0-856ac7158fa4 service nova] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Received event network-vif-deleted-8c01ddea-32d5-4b65-9774-46a15e7c391f {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 593.828953] env[61728]: DEBUG nova.compute.utils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 593.830506] env[61728]: DEBUG nova.compute.manager [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 593.830681] env[61728]: DEBUG nova.network.neutron [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 593.871866] env[61728]: DEBUG nova.policy [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '091abd20ebfe450b9283f5c7df1c7993', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c78292171414530a1caba3dfd0f620f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 594.138769] env[61728]: DEBUG nova.network.neutron [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Successfully created port: 7a9a7fa8-9c61-46be-a49a-8dd67bce2d22 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 594.182355] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524d54cf-b512-2e26-9ec0-c4e7fbc48e7e, 'name': SearchDatastore_Task, 'duration_secs': 0.03089} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.185783] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.186057] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 594.186303] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.186507] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.186622] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 594.186888] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f40d703-8cfd-49f4-a4d3-37557693eadd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.195741] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-463663, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067853} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.198831] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 594.198831] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 594.198831] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 594.198831] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41faee39-62b7-4299-8454-9c2c540d60ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.201524] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dffcfdde-65ca-4ca6-a727-f6c4096e8237 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.206896] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 594.206896] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528662b9-6e73-bc19-1f10-f09a4170fef7" [ 594.206896] env[61728]: _type = "Task" [ 594.206896] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.225367] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Reconfiguring VM instance instance-00000017 to attach disk [datastore1] 4e5dc620-68a1-4e10-a8be-702c4999ca10/4e5dc620-68a1-4e10-a8be-702c4999ca10.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 594.229630] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e20dbe59-1c33-4c3b-acf8-3816b05c8781 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.252254] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528662b9-6e73-bc19-1f10-f09a4170fef7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.253623] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Waiting for the task: (returnval){ [ 594.253623] env[61728]: value = "task-463664" [ 594.253623] env[61728]: _type = "Task" [ 594.253623] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.262996] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-463664, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.334062] env[61728]: DEBUG nova.compute.manager [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 594.467762] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Acquiring lock "2c76f45a-b87c-4c3b-9bf9-ba5d67265437" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.468026] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Lock "2c76f45a-b87c-4c3b-9bf9-ba5d67265437" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.468234] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Acquiring lock "2c76f45a-b87c-4c3b-9bf9-ba5d67265437-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.468455] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Lock "2c76f45a-b87c-4c3b-9bf9-ba5d67265437-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.468648] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Lock "2c76f45a-b87c-4c3b-9bf9-ba5d67265437-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.470631] env[61728]: INFO nova.compute.manager [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Terminating instance [ 594.474500] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Acquiring lock "refresh_cache-2c76f45a-b87c-4c3b-9bf9-ba5d67265437" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.474716] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Acquired lock "refresh_cache-2c76f45a-b87c-4c3b-9bf9-ba5d67265437" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.474897] env[61728]: DEBUG nova.network.neutron [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 594.727020] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528662b9-6e73-bc19-1f10-f09a4170fef7, 'name': SearchDatastore_Task, 'duration_secs': 0.06632} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.729708] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f26a149a-d13f-41d1-858f-2e97f3f78313 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.734638] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 594.734638] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5200c1f6-8032-e337-03bd-8dc0ef572c1f" [ 594.734638] env[61728]: _type = "Task" [ 594.734638] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.742267] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5200c1f6-8032-e337-03bd-8dc0ef572c1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.763210] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-463664, 'name': ReconfigVM_Task, 'duration_secs': 0.422583} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.763481] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Reconfigured VM instance instance-00000017 to attach disk [datastore1] 4e5dc620-68a1-4e10-a8be-702c4999ca10/4e5dc620-68a1-4e10-a8be-702c4999ca10.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 594.764125] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-843350ed-6781-45be-93e1-150ea007b26b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.772992] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Waiting for the task: (returnval){ [ 594.772992] env[61728]: value = "task-463665" [ 594.772992] env[61728]: _type = "Task" [ 594.772992] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.782110] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-463665, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.794682] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a8ceb1-ab51-488a-97d2-c6c386a7f058 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.801996] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1f4c02-64ab-488e-8ddc-b94084386cc5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.831772] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f46a70f-cbb4-48a2-8889-ae1090114ab8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.839665] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1148487-3554-49f3-b263-0de107c055b1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.858503] env[61728]: DEBUG nova.compute.provider_tree [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.154993] env[61728]: DEBUG nova.network.neutron [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.220910] env[61728]: DEBUG nova.network.neutron [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.246964] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5200c1f6-8032-e337-03bd-8dc0ef572c1f, 'name': SearchDatastore_Task, 'duration_secs': 0.009336} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.247234] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.247478] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2/7b16fbc0-7f13-405f-b84e-e18de1ca7dd2.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 595.247959] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1e037426-7184-4e5f-9967-e7250e8e0c91 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.254653] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 595.254653] env[61728]: value = "task-463666" [ 595.254653] env[61728]: _type = "Task" [ 595.254653] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.262371] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463666, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.281612] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-463665, 'name': Rename_Task, 'duration_secs': 0.133651} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.281899] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 595.282181] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0cf56876-2e04-4a29-8b21-8958c1307d1f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.288348] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Waiting for the task: (returnval){ [ 595.288348] env[61728]: value = "task-463667" [ 595.288348] env[61728]: _type = "Task" [ 595.288348] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.297983] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-463667, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.349703] env[61728]: DEBUG nova.compute.manager [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 595.362090] env[61728]: DEBUG nova.scheduler.client.report [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.381394] env[61728]: DEBUG nova.virt.hardware [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 595.381672] env[61728]: DEBUG nova.virt.hardware [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 595.382558] env[61728]: DEBUG nova.virt.hardware [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 595.382558] env[61728]: DEBUG nova.virt.hardware [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 595.382558] env[61728]: DEBUG nova.virt.hardware [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 595.382558] env[61728]: DEBUG nova.virt.hardware [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 595.382997] env[61728]: DEBUG nova.virt.hardware [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 595.383183] env[61728]: DEBUG nova.virt.hardware [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 595.383392] env[61728]: DEBUG nova.virt.hardware [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 595.383572] env[61728]: DEBUG nova.virt.hardware [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 595.383742] env[61728]: DEBUG nova.virt.hardware [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 595.384654] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39932a32-196f-4ed7-8d6e-c24179b7eb11 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.395712] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1844ca8d-5743-47da-9c07-6eff3c7f3d5c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.667280] env[61728]: DEBUG nova.network.neutron [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Successfully updated port: 7a9a7fa8-9c61-46be-a49a-8dd67bce2d22 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 595.687347] env[61728]: DEBUG nova.compute.manager [req-cd4b4625-31d3-408e-8971-f73c5c29942c req-9a681816-8bc7-4bb1-a257-7f8d04f1077f service nova] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Received event network-vif-plugged-7a9a7fa8-9c61-46be-a49a-8dd67bce2d22 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 595.687347] env[61728]: DEBUG oslo_concurrency.lockutils [req-cd4b4625-31d3-408e-8971-f73c5c29942c req-9a681816-8bc7-4bb1-a257-7f8d04f1077f service nova] Acquiring lock "fc3c1d93-b095-4d80-8d71-243b66d85a46-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.688026] env[61728]: DEBUG oslo_concurrency.lockutils [req-cd4b4625-31d3-408e-8971-f73c5c29942c req-9a681816-8bc7-4bb1-a257-7f8d04f1077f service nova] Lock "fc3c1d93-b095-4d80-8d71-243b66d85a46-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.688026] env[61728]: DEBUG oslo_concurrency.lockutils [req-cd4b4625-31d3-408e-8971-f73c5c29942c req-9a681816-8bc7-4bb1-a257-7f8d04f1077f service nova] Lock "fc3c1d93-b095-4d80-8d71-243b66d85a46-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.688026] env[61728]: DEBUG nova.compute.manager [req-cd4b4625-31d3-408e-8971-f73c5c29942c req-9a681816-8bc7-4bb1-a257-7f8d04f1077f service nova] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] No waiting events found dispatching network-vif-plugged-7a9a7fa8-9c61-46be-a49a-8dd67bce2d22 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 595.688200] env[61728]: WARNING nova.compute.manager [req-cd4b4625-31d3-408e-8971-f73c5c29942c req-9a681816-8bc7-4bb1-a257-7f8d04f1077f service nova] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Received unexpected event network-vif-plugged-7a9a7fa8-9c61-46be-a49a-8dd67bce2d22 for instance with vm_state building and task_state spawning. [ 595.723988] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Releasing lock "refresh_cache-2c76f45a-b87c-4c3b-9bf9-ba5d67265437" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.724341] env[61728]: DEBUG nova.compute.manager [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 595.724554] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 595.725485] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb75ce0a-9401-40fc-9f04-b05cf32f5fe3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.733166] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 595.733314] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f85a0e01-44cd-47d6-a5a9-d4a3842d2433 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.740079] env[61728]: DEBUG oslo_vmware.api [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Waiting for the task: (returnval){ [ 595.740079] env[61728]: value = "task-463668" [ 595.740079] env[61728]: _type = "Task" [ 595.740079] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.748895] env[61728]: DEBUG oslo_vmware.api [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463668, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.765096] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463666, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45156} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.765547] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2/7b16fbc0-7f13-405f-b84e-e18de1ca7dd2.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 595.765818] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 595.766135] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d76af79-9eb4-4a53-befd-aa1d81d0174e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.773233] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 595.773233] env[61728]: value = "task-463669" [ 595.773233] env[61728]: _type = "Task" [ 595.773233] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.782087] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463669, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.797372] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-463667, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.867849] env[61728]: DEBUG oslo_concurrency.lockutils [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.539s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.869034] env[61728]: DEBUG nova.compute.manager [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.871196] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 31.329s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.170957] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "refresh_cache-fc3c1d93-b095-4d80-8d71-243b66d85a46" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.171138] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "refresh_cache-fc3c1d93-b095-4d80-8d71-243b66d85a46" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.171248] env[61728]: DEBUG nova.network.neutron [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 596.253201] env[61728]: DEBUG oslo_vmware.api [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463668, 'name': PowerOffVM_Task, 'duration_secs': 0.1302} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.253540] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 596.253772] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 596.254087] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-33b00b73-ff56-4952-9418-a5ca401aec3d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.281895] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 596.282170] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 596.282454] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Deleting the datastore file [datastore1] 2c76f45a-b87c-4c3b-9bf9-ba5d67265437 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 596.283131] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5ba91134-f96c-4ea5-bf53-f8fb38ced0ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.288857] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463669, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067724} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.289540] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 596.290620] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf76041-7493-456a-b726-319b4e26f587 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.298502] env[61728]: DEBUG oslo_vmware.api [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Waiting for the task: (returnval){ [ 596.298502] env[61728]: value = "task-463671" [ 596.298502] env[61728]: _type = "Task" [ 596.298502] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.333577] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Reconfiguring VM instance instance-00000018 to attach disk [datastore1] 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2/7b16fbc0-7f13-405f-b84e-e18de1ca7dd2.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 596.333984] env[61728]: DEBUG oslo_vmware.api [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-463667, 'name': PowerOnVM_Task, 'duration_secs': 0.840061} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.334766] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff978845-d5a9-4d76-93e7-98a698e8937c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.356605] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 596.356783] env[61728]: INFO nova.compute.manager [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Took 8.53 seconds to spawn the instance on the hypervisor. [ 596.357037] env[61728]: DEBUG nova.compute.manager [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 596.361983] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c949bbca-ab7a-4fe0-9dbf-f4552c54e565 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.365710] env[61728]: DEBUG oslo_vmware.api [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463671, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.378985] env[61728]: DEBUG nova.compute.utils [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.380266] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 596.380266] env[61728]: value = "task-463672" [ 596.380266] env[61728]: _type = "Task" [ 596.380266] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.381613] env[61728]: DEBUG nova.compute.manager [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 596.381613] env[61728]: DEBUG nova.network.neutron [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 596.391801] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463672, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.443492] env[61728]: DEBUG nova.policy [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '659d9344c72e4458938b15ef286fa78b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c1aaff1dd21b497eba4d4dc9643534db', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 596.732075] env[61728]: DEBUG nova.network.neutron [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.764964] env[61728]: DEBUG nova.network.neutron [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Successfully created port: b797bbf1-77c6-48d3-b63c-f4caf481f773 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 596.813688] env[61728]: DEBUG oslo_vmware.api [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Task: {'id': task-463671, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228822} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.813933] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 596.814126] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 596.814298] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 596.814883] env[61728]: INFO nova.compute.manager [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Took 1.09 seconds to destroy the instance on the hypervisor. [ 596.814883] env[61728]: DEBUG oslo.service.loopingcall [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 596.815086] env[61728]: DEBUG nova.compute.manager [-] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.815086] env[61728]: DEBUG nova.network.neutron [-] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 596.837890] env[61728]: DEBUG nova.network.neutron [-] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.884196] env[61728]: DEBUG nova.compute.manager [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.900573] env[61728]: INFO nova.compute.manager [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Took 46.98 seconds to build instance. [ 596.911019] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463672, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.948330] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5a2797-00b1-4f90-a415-09fd5dbcde37 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.959125] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e180785-bfa8-437c-8366-8e10ba58574a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.992424] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb4bec2-d432-467f-82f5-7c91b5ed2d3e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.001056] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b9a8a0-48df-42c5-9415-cae0722c8565 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.014889] env[61728]: DEBUG nova.compute.provider_tree [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.023401] env[61728]: DEBUG nova.network.neutron [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Updating instance_info_cache with network_info: [{"id": "7a9a7fa8-9c61-46be-a49a-8dd67bce2d22", "address": "fa:16:3e:22:cc:f7", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a9a7fa8-9c", "ovs_interfaceid": "7a9a7fa8-9c61-46be-a49a-8dd67bce2d22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.346693] env[61728]: DEBUG nova.network.neutron [-] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.397658] env[61728]: INFO nova.virt.block_device [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Booting with volume 2e85ae80-363c-4fa7-add2-31dd95e23cfd at /dev/sda [ 597.403284] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03afe0eb-7f35-41c5-8a24-256ba8f3b294 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Lock "4e5dc620-68a1-4e10-a8be-702c4999ca10" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.151s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.420789] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463672, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.450099] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-786e788a-851a-4e7e-a1be-ab41336c3a6b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.461605] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc7c74e-d42d-46a3-b0e4-b30e540c5e4c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.498519] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8bd74917-3b35-45e2-8ad1-2296b6ad951e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.507918] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bde6bf7-7cc8-4963-9e4e-a59d2cc365ac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.521026] env[61728]: DEBUG nova.scheduler.client.report [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.527320] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "refresh_cache-fc3c1d93-b095-4d80-8d71-243b66d85a46" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.527660] env[61728]: DEBUG nova.compute.manager [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Instance network_info: |[{"id": "7a9a7fa8-9c61-46be-a49a-8dd67bce2d22", "address": "fa:16:3e:22:cc:f7", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a9a7fa8-9c", "ovs_interfaceid": "7a9a7fa8-9c61-46be-a49a-8dd67bce2d22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 597.528084] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:cc:f7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7a9a7fa8-9c61-46be-a49a-8dd67bce2d22', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 597.538366] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Creating folder: Project (0c78292171414530a1caba3dfd0f620f). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 597.552416] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ee7ac390-0123-4aff-a635-955eeeb466e0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.555675] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad55e0f-7fd3-434f-8fd8-04be3b0bd20d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.563682] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef892c9-6024-4490-8b5b-79ea4cd68898 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.567758] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Created folder: Project (0c78292171414530a1caba3dfd0f620f) in parent group-v121913. [ 597.567951] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Creating folder: Instances. Parent ref: group-v121991. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 597.568674] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd325106-641d-454e-be89-4a19923bd5bf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.579394] env[61728]: DEBUG nova.virt.block_device [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Updating existing volume attachment record: 284f10ec-d088-409f-a18d-8965e1ac685d {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 597.583984] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Created folder: Instances in parent group-v121991. [ 597.584157] env[61728]: DEBUG oslo.service.loopingcall [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 597.584347] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 597.584622] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63cb3b3b-5b50-49e1-80d5-d0f30f12b6a2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.605196] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 597.605196] env[61728]: value = "task-463675" [ 597.605196] env[61728]: _type = "Task" [ 597.605196] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.615509] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463675, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.818954] env[61728]: DEBUG nova.compute.manager [req-9aa6348a-6588-4744-8918-a5ba2237af3c req-5ebca182-2c56-4e11-8892-4b8434a77582 service nova] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Received event network-changed-7a9a7fa8-9c61-46be-a49a-8dd67bce2d22 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 597.818954] env[61728]: DEBUG nova.compute.manager [req-9aa6348a-6588-4744-8918-a5ba2237af3c req-5ebca182-2c56-4e11-8892-4b8434a77582 service nova] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Refreshing instance network info cache due to event network-changed-7a9a7fa8-9c61-46be-a49a-8dd67bce2d22. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 597.818954] env[61728]: DEBUG oslo_concurrency.lockutils [req-9aa6348a-6588-4744-8918-a5ba2237af3c req-5ebca182-2c56-4e11-8892-4b8434a77582 service nova] Acquiring lock "refresh_cache-fc3c1d93-b095-4d80-8d71-243b66d85a46" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.818954] env[61728]: DEBUG oslo_concurrency.lockutils [req-9aa6348a-6588-4744-8918-a5ba2237af3c req-5ebca182-2c56-4e11-8892-4b8434a77582 service nova] Acquired lock "refresh_cache-fc3c1d93-b095-4d80-8d71-243b66d85a46" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.819486] env[61728]: DEBUG nova.network.neutron [req-9aa6348a-6588-4744-8918-a5ba2237af3c req-5ebca182-2c56-4e11-8892-4b8434a77582 service nova] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Refreshing network info cache for port 7a9a7fa8-9c61-46be-a49a-8dd67bce2d22 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 597.853172] env[61728]: INFO nova.compute.manager [-] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Took 1.04 seconds to deallocate network for instance. [ 597.913491] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463672, 'name': ReconfigVM_Task, 'duration_secs': 1.158193} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.913491] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Reconfigured VM instance instance-00000018 to attach disk [datastore1] 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2/7b16fbc0-7f13-405f-b84e-e18de1ca7dd2.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 597.913491] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69c75147-5946-4b88-a009-9b6fb9adf850 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.915494] env[61728]: DEBUG nova.compute.manager [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.919849] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 597.919849] env[61728]: value = "task-463676" [ 597.919849] env[61728]: _type = "Task" [ 597.919849] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.933735] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463676, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.095766] env[61728]: DEBUG nova.compute.manager [req-c6b0bbfb-e1ce-4a37-9e5e-26c906003abb req-ab082080-461a-4a4f-ab8c-007b11042585 service nova] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Received event network-changed-6408b220-22b3-4b0d-961e-9a7a9ae42648 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 598.095958] env[61728]: DEBUG nova.compute.manager [req-c6b0bbfb-e1ce-4a37-9e5e-26c906003abb req-ab082080-461a-4a4f-ab8c-007b11042585 service nova] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Refreshing instance network info cache due to event network-changed-6408b220-22b3-4b0d-961e-9a7a9ae42648. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 598.096200] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6b0bbfb-e1ce-4a37-9e5e-26c906003abb req-ab082080-461a-4a4f-ab8c-007b11042585 service nova] Acquiring lock "refresh_cache-4e5dc620-68a1-4e10-a8be-702c4999ca10" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.096345] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6b0bbfb-e1ce-4a37-9e5e-26c906003abb req-ab082080-461a-4a4f-ab8c-007b11042585 service nova] Acquired lock "refresh_cache-4e5dc620-68a1-4e10-a8be-702c4999ca10" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.096504] env[61728]: DEBUG nova.network.neutron [req-c6b0bbfb-e1ce-4a37-9e5e-26c906003abb req-ab082080-461a-4a4f-ab8c-007b11042585 service nova] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Refreshing network info cache for port 6408b220-22b3-4b0d-961e-9a7a9ae42648 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 598.120029] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463675, 'name': CreateVM_Task, 'duration_secs': 0.39513} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.120029] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 598.120029] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.120029] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.120029] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 598.120029] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-402ac375-6fb4-4b8a-828f-e6b3ede898fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.125885] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 598.125885] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5263959f-d9ad-7a49-45d1-85943f6ff539" [ 598.125885] env[61728]: _type = "Task" [ 598.125885] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.132024] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5263959f-d9ad-7a49-45d1-85943f6ff539, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.361863] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.436614] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463676, 'name': Rename_Task, 'duration_secs': 0.146171} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.437536] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 598.444176] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d233048-9f1d-4d5e-b118-94edd4c0b322 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.447200] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.452563] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 598.452563] env[61728]: value = "task-463677" [ 598.452563] env[61728]: _type = "Task" [ 598.452563] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.463868] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463677, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.463868] env[61728]: DEBUG nova.network.neutron [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Successfully updated port: b797bbf1-77c6-48d3-b63c-f4caf481f773 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 598.533026] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.662s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.536196] env[61728]: DEBUG oslo_concurrency.lockutils [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.891s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.536380] env[61728]: DEBUG nova.objects.instance [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61728) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 598.639324] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5263959f-d9ad-7a49-45d1-85943f6ff539, 'name': SearchDatastore_Task, 'duration_secs': 0.009751} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.639708] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.639976] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 598.640277] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.640445] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.640631] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 598.640888] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-496a3e81-b0ca-4777-800e-5b0837510c56 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.650532] env[61728]: DEBUG nova.network.neutron [req-9aa6348a-6588-4744-8918-a5ba2237af3c req-5ebca182-2c56-4e11-8892-4b8434a77582 service nova] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Updated VIF entry in instance network info cache for port 7a9a7fa8-9c61-46be-a49a-8dd67bce2d22. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 598.650869] env[61728]: DEBUG nova.network.neutron [req-9aa6348a-6588-4744-8918-a5ba2237af3c req-5ebca182-2c56-4e11-8892-4b8434a77582 service nova] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Updating instance_info_cache with network_info: [{"id": "7a9a7fa8-9c61-46be-a49a-8dd67bce2d22", "address": "fa:16:3e:22:cc:f7", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a9a7fa8-9c", "ovs_interfaceid": "7a9a7fa8-9c61-46be-a49a-8dd67bce2d22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.654796] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 598.655257] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 598.658137] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-935a32a0-198f-4cd8-8d93-12facbebd93a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.663884] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 598.663884] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52088df0-b219-a000-8914-ea6bf516bad3" [ 598.663884] env[61728]: _type = "Task" [ 598.663884] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.676417] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52088df0-b219-a000-8914-ea6bf516bad3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.840796] env[61728]: DEBUG nova.network.neutron [req-c6b0bbfb-e1ce-4a37-9e5e-26c906003abb req-ab082080-461a-4a4f-ab8c-007b11042585 service nova] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Updated VIF entry in instance network info cache for port 6408b220-22b3-4b0d-961e-9a7a9ae42648. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 598.841213] env[61728]: DEBUG nova.network.neutron [req-c6b0bbfb-e1ce-4a37-9e5e-26c906003abb req-ab082080-461a-4a4f-ab8c-007b11042585 service nova] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Updating instance_info_cache with network_info: [{"id": "6408b220-22b3-4b0d-961e-9a7a9ae42648", "address": "fa:16:3e:31:7a:9d", "network": {"id": "36cd4339-1796-4b57-85f8-950073dcd063", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1977850529-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62e6ae50d2a34b7280e409111d327a0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6408b220-22", "ovs_interfaceid": "6408b220-22b3-4b0d-961e-9a7a9ae42648", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.962887] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463677, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.966607] env[61728]: DEBUG oslo_concurrency.lockutils [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Acquiring lock "refresh_cache-37b8ca5a-e0f8-414b-a363-e56db520f027" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.966760] env[61728]: DEBUG oslo_concurrency.lockutils [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Acquired lock "refresh_cache-37b8ca5a-e0f8-414b-a363-e56db520f027" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.966916] env[61728]: DEBUG nova.network.neutron [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 599.104692] env[61728]: INFO nova.scheduler.client.report [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Deleted allocation for migration 04adf437-6f01-45eb-848e-3da8ff14cd18 [ 599.156925] env[61728]: DEBUG oslo_concurrency.lockutils [req-9aa6348a-6588-4744-8918-a5ba2237af3c req-5ebca182-2c56-4e11-8892-4b8434a77582 service nova] Releasing lock "refresh_cache-fc3c1d93-b095-4d80-8d71-243b66d85a46" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.176544] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52088df0-b219-a000-8914-ea6bf516bad3, 'name': SearchDatastore_Task, 'duration_secs': 0.011511} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.177374] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed72db44-c00c-4439-aa47-8037ddaef206 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.183127] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 599.183127] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f86106-a725-75b3-0c98-2a502607284b" [ 599.183127] env[61728]: _type = "Task" [ 599.183127] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.190919] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f86106-a725-75b3-0c98-2a502607284b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.344260] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6b0bbfb-e1ce-4a37-9e5e-26c906003abb req-ab082080-461a-4a4f-ab8c-007b11042585 service nova] Releasing lock "refresh_cache-4e5dc620-68a1-4e10-a8be-702c4999ca10" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.472893] env[61728]: DEBUG oslo_vmware.api [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463677, 'name': PowerOnVM_Task, 'duration_secs': 0.654074} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.473186] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 599.473387] env[61728]: INFO nova.compute.manager [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Took 9.01 seconds to spawn the instance on the hypervisor. [ 599.473690] env[61728]: DEBUG nova.compute.manager [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 599.475274] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6039cef9-2953-462e-877f-a55e443e5a24 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.509521] env[61728]: DEBUG nova.network.neutron [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.553727] env[61728]: DEBUG oslo_concurrency.lockutils [None req-96cb05ac-83b3-457d-b7de-0f6a17a80a80 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.554730] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.468s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.556765] env[61728]: INFO nova.compute.claims [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.610784] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f8080628-0242-41e6-9c2f-3fe0a82d12ce tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 38.524s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.672030] env[61728]: DEBUG nova.network.neutron [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Updating instance_info_cache with network_info: [{"id": "b797bbf1-77c6-48d3-b63c-f4caf481f773", "address": "fa:16:3e:ed:af:31", "network": {"id": "cd3b3b4d-d9ae-4037-825a-5693706df99c", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-2112414125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1aaff1dd21b497eba4d4dc9643534db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c8a5d7c-ee1f-4a41-94e4-db31e85a398d", "external-id": "cl2-zone-613", "segmentation_id": 613, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb797bbf1-77", "ovs_interfaceid": "b797bbf1-77c6-48d3-b63c-f4caf481f773", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.696064] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f86106-a725-75b3-0c98-2a502607284b, 'name': SearchDatastore_Task, 'duration_secs': 0.013184} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.696232] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.696479] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] fc3c1d93-b095-4d80-8d71-243b66d85a46/fc3c1d93-b095-4d80-8d71-243b66d85a46.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 599.696729] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ff9ce7e5-8505-456e-bbe2-2dbbef409377 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.703412] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 599.703412] env[61728]: value = "task-463678" [ 599.703412] env[61728]: _type = "Task" [ 599.703412] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.711624] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463678, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.738674] env[61728]: DEBUG nova.compute.manager [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 599.740208] env[61728]: DEBUG nova.virt.hardware [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 599.740208] env[61728]: DEBUG nova.virt.hardware [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 599.740208] env[61728]: DEBUG nova.virt.hardware [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 599.740208] env[61728]: DEBUG nova.virt.hardware [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 599.740616] env[61728]: DEBUG nova.virt.hardware [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 599.740616] env[61728]: DEBUG nova.virt.hardware [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 599.740616] env[61728]: DEBUG nova.virt.hardware [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 599.740616] env[61728]: DEBUG nova.virt.hardware [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 599.740763] env[61728]: DEBUG nova.virt.hardware [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 599.740934] env[61728]: DEBUG nova.virt.hardware [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 599.741297] env[61728]: DEBUG nova.virt.hardware [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 599.742383] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ebbfa1-82f5-4456-8e60-dfa3d4c4f9e3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.754440] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3098b073-cb1e-4ac4-b232-88717323727c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.845424] env[61728]: DEBUG nova.compute.manager [req-64cd488c-cea2-4b2f-a45f-5cc81517d76d req-97dbc425-12b5-4b36-b90e-949301f028af service nova] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Received event network-vif-plugged-b797bbf1-77c6-48d3-b63c-f4caf481f773 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 599.845698] env[61728]: DEBUG oslo_concurrency.lockutils [req-64cd488c-cea2-4b2f-a45f-5cc81517d76d req-97dbc425-12b5-4b36-b90e-949301f028af service nova] Acquiring lock "37b8ca5a-e0f8-414b-a363-e56db520f027-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.845938] env[61728]: DEBUG oslo_concurrency.lockutils [req-64cd488c-cea2-4b2f-a45f-5cc81517d76d req-97dbc425-12b5-4b36-b90e-949301f028af service nova] Lock "37b8ca5a-e0f8-414b-a363-e56db520f027-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.846161] env[61728]: DEBUG oslo_concurrency.lockutils [req-64cd488c-cea2-4b2f-a45f-5cc81517d76d req-97dbc425-12b5-4b36-b90e-949301f028af service nova] Lock "37b8ca5a-e0f8-414b-a363-e56db520f027-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.846445] env[61728]: DEBUG nova.compute.manager [req-64cd488c-cea2-4b2f-a45f-5cc81517d76d req-97dbc425-12b5-4b36-b90e-949301f028af service nova] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] No waiting events found dispatching network-vif-plugged-b797bbf1-77c6-48d3-b63c-f4caf481f773 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 599.846657] env[61728]: WARNING nova.compute.manager [req-64cd488c-cea2-4b2f-a45f-5cc81517d76d req-97dbc425-12b5-4b36-b90e-949301f028af service nova] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Received unexpected event network-vif-plugged-b797bbf1-77c6-48d3-b63c-f4caf481f773 for instance with vm_state building and task_state spawning. [ 599.846878] env[61728]: DEBUG nova.compute.manager [req-64cd488c-cea2-4b2f-a45f-5cc81517d76d req-97dbc425-12b5-4b36-b90e-949301f028af service nova] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Received event network-changed-b797bbf1-77c6-48d3-b63c-f4caf481f773 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 599.847060] env[61728]: DEBUG nova.compute.manager [req-64cd488c-cea2-4b2f-a45f-5cc81517d76d req-97dbc425-12b5-4b36-b90e-949301f028af service nova] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Refreshing instance network info cache due to event network-changed-b797bbf1-77c6-48d3-b63c-f4caf481f773. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 599.847279] env[61728]: DEBUG oslo_concurrency.lockutils [req-64cd488c-cea2-4b2f-a45f-5cc81517d76d req-97dbc425-12b5-4b36-b90e-949301f028af service nova] Acquiring lock "refresh_cache-37b8ca5a-e0f8-414b-a363-e56db520f027" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.996411] env[61728]: INFO nova.compute.manager [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Took 46.18 seconds to build instance. [ 600.174613] env[61728]: DEBUG oslo_concurrency.lockutils [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Releasing lock "refresh_cache-37b8ca5a-e0f8-414b-a363-e56db520f027" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.175113] env[61728]: DEBUG nova.compute.manager [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Instance network_info: |[{"id": "b797bbf1-77c6-48d3-b63c-f4caf481f773", "address": "fa:16:3e:ed:af:31", "network": {"id": "cd3b3b4d-d9ae-4037-825a-5693706df99c", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-2112414125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1aaff1dd21b497eba4d4dc9643534db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c8a5d7c-ee1f-4a41-94e4-db31e85a398d", "external-id": "cl2-zone-613", "segmentation_id": 613, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb797bbf1-77", "ovs_interfaceid": "b797bbf1-77c6-48d3-b63c-f4caf481f773", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 600.175449] env[61728]: DEBUG oslo_concurrency.lockutils [req-64cd488c-cea2-4b2f-a45f-5cc81517d76d req-97dbc425-12b5-4b36-b90e-949301f028af service nova] Acquired lock "refresh_cache-37b8ca5a-e0f8-414b-a363-e56db520f027" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.175686] env[61728]: DEBUG nova.network.neutron [req-64cd488c-cea2-4b2f-a45f-5cc81517d76d req-97dbc425-12b5-4b36-b90e-949301f028af service nova] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Refreshing network info cache for port b797bbf1-77c6-48d3-b63c-f4caf481f773 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 600.177264] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ed:af:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c8a5d7c-ee1f-4a41-94e4-db31e85a398d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b797bbf1-77c6-48d3-b63c-f4caf481f773', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 600.185417] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Creating folder: Project (c1aaff1dd21b497eba4d4dc9643534db). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 600.188739] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0071a1e8-ef2a-499a-8db8-d2016dac5155 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.202847] env[61728]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 600.203126] env[61728]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61728) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 600.203617] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Folder already exists: Project (c1aaff1dd21b497eba4d4dc9643534db). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 600.203878] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Creating folder: Instances. Parent ref: group-v121929. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 600.204255] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9ac86d4e-8c5c-4980-855a-7049a3fd1f36 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.215277] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463678, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.219063] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Created folder: Instances in parent group-v121929. [ 600.219063] env[61728]: DEBUG oslo.service.loopingcall [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 600.219063] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 600.219063] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-170054a6-b5ee-4e99-8a58-aca6f3cba50c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.241891] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 600.241891] env[61728]: value = "task-463681" [ 600.241891] env[61728]: _type = "Task" [ 600.241891] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.249959] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463681, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.436372] env[61728]: DEBUG nova.network.neutron [req-64cd488c-cea2-4b2f-a45f-5cc81517d76d req-97dbc425-12b5-4b36-b90e-949301f028af service nova] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Updated VIF entry in instance network info cache for port b797bbf1-77c6-48d3-b63c-f4caf481f773. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 600.436734] env[61728]: DEBUG nova.network.neutron [req-64cd488c-cea2-4b2f-a45f-5cc81517d76d req-97dbc425-12b5-4b36-b90e-949301f028af service nova] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Updating instance_info_cache with network_info: [{"id": "b797bbf1-77c6-48d3-b63c-f4caf481f773", "address": "fa:16:3e:ed:af:31", "network": {"id": "cd3b3b4d-d9ae-4037-825a-5693706df99c", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-2112414125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1aaff1dd21b497eba4d4dc9643534db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c8a5d7c-ee1f-4a41-94e4-db31e85a398d", "external-id": "cl2-zone-613", "segmentation_id": 613, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb797bbf1-77", "ovs_interfaceid": "b797bbf1-77c6-48d3-b63c-f4caf481f773", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.500921] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c811c834-1ed7-45a2-b4d7-9e661da061d3 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.842s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.714431] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463678, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.752330] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463681, 'name': CreateVM_Task, 'duration_secs': 0.39592} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.754838] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 600.756047] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': '284f10ec-d088-409f-a18d-8965e1ac685d', 'delete_on_termination': True, 'guest_format': None, 'disk_bus': None, 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-121945', 'volume_id': '2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'name': 'volume-2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '37b8ca5a-e0f8-414b-a363-e56db520f027', 'attached_at': '', 'detached_at': '', 'volume_id': '2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'serial': '2e85ae80-363c-4fa7-add2-31dd95e23cfd'}, 'mount_device': '/dev/sda', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=61728) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 600.756047] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Root volume attach. Driver type: vmdk {{(pid=61728) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 600.756689] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e9fe745-7142-43cb-8a14-d25b0355e0eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.764209] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-718cad6e-72fb-45df-9d60-75c1f2e8703d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.774373] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cacc8e2e-e884-4b3b-94fd-18d2f24e5338 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.782539] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-b31e3cc8-ccfb-4848-baed-c393a4cc4c92 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.790322] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Waiting for the task: (returnval){ [ 600.790322] env[61728]: value = "task-463682" [ 600.790322] env[61728]: _type = "Task" [ 600.790322] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.798224] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-463682, 'name': RelocateVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.940212] env[61728]: DEBUG oslo_concurrency.lockutils [req-64cd488c-cea2-4b2f-a45f-5cc81517d76d req-97dbc425-12b5-4b36-b90e-949301f028af service nova] Releasing lock "refresh_cache-37b8ca5a-e0f8-414b-a363-e56db520f027" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.005258] env[61728]: DEBUG nova.compute.manager [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 601.072034] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec64803-9e26-4968-81d5-bd01eab12a7b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.080617] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a2a904-506c-422b-80b2-36ee5bfebd98 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.115753] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d4a80b-5d0f-47ad-9403-aa68e5ef2a89 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.127244] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2547bdab-5985-48aa-b4a3-0e8924aa4638 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.138598] env[61728]: DEBUG nova.compute.provider_tree [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.201772] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "1d9114cb-a0af-4874-962f-27237b3c89cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.202194] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "1d9114cb-a0af-4874-962f-27237b3c89cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.215522] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463678, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.303324] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-463682, 'name': RelocateVM_Task, 'duration_secs': 0.453313} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.303324] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 601.303324] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-121945', 'volume_id': '2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'name': 'volume-2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '37b8ca5a-e0f8-414b-a363-e56db520f027', 'attached_at': '', 'detached_at': '', 'volume_id': '2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'serial': '2e85ae80-363c-4fa7-add2-31dd95e23cfd'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 601.303324] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4259d6f9-c765-4e69-b914-8b14cf0e28a8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.318298] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3efb32e2-bee5-45b2-be4a-ebfe64b7b96b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.342087] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Reconfiguring VM instance instance-0000001a to attach disk [datastore1] volume-2e85ae80-363c-4fa7-add2-31dd95e23cfd/volume-2e85ae80-363c-4fa7-add2-31dd95e23cfd.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 601.342557] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5690d05-5fad-479a-91e9-e857fe0033e1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.362558] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Waiting for the task: (returnval){ [ 601.362558] env[61728]: value = "task-463683" [ 601.362558] env[61728]: _type = "Task" [ 601.362558] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.371519] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-463683, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.522710] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.645026] env[61728]: DEBUG nova.scheduler.client.report [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.717267] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463678, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.830029} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.718497] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] fc3c1d93-b095-4d80-8d71-243b66d85a46/fc3c1d93-b095-4d80-8d71-243b66d85a46.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 601.721023] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 601.721023] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f7d25db-8579-459e-8b0c-74b2c9c4005e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.728027] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 601.728027] env[61728]: value = "task-463684" [ 601.728027] env[61728]: _type = "Task" [ 601.728027] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.740973] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463684, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.874022] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-463683, 'name': ReconfigVM_Task, 'duration_secs': 0.326049} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.874678] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Reconfigured VM instance instance-0000001a to attach disk [datastore1] volume-2e85ae80-363c-4fa7-add2-31dd95e23cfd/volume-2e85ae80-363c-4fa7-add2-31dd95e23cfd.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 601.881154] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f98cc9a8-c20a-4f12-a7ba-efaf12b2fd25 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.901083] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Waiting for the task: (returnval){ [ 601.901083] env[61728]: value = "task-463685" [ 601.901083] env[61728]: _type = "Task" [ 601.901083] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.909574] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-463685, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.007078] env[61728]: DEBUG nova.compute.manager [req-6248bfbf-1fe2-4306-8e27-37d2646f157e req-fb46ada2-2f77-431a-9d1c-9c9d0978c06f service nova] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Received event network-changed-6c9a6b61-5248-4996-a390-c18edb3ec021 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 602.007367] env[61728]: DEBUG nova.compute.manager [req-6248bfbf-1fe2-4306-8e27-37d2646f157e req-fb46ada2-2f77-431a-9d1c-9c9d0978c06f service nova] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Refreshing instance network info cache due to event network-changed-6c9a6b61-5248-4996-a390-c18edb3ec021. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 602.007653] env[61728]: DEBUG oslo_concurrency.lockutils [req-6248bfbf-1fe2-4306-8e27-37d2646f157e req-fb46ada2-2f77-431a-9d1c-9c9d0978c06f service nova] Acquiring lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.007706] env[61728]: DEBUG oslo_concurrency.lockutils [req-6248bfbf-1fe2-4306-8e27-37d2646f157e req-fb46ada2-2f77-431a-9d1c-9c9d0978c06f service nova] Acquired lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.007861] env[61728]: DEBUG nova.network.neutron [req-6248bfbf-1fe2-4306-8e27-37d2646f157e req-fb46ada2-2f77-431a-9d1c-9c9d0978c06f service nova] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Refreshing network info cache for port 6c9a6b61-5248-4996-a390-c18edb3ec021 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 602.148498] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.594s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.149010] env[61728]: DEBUG nova.compute.manager [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.152521] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.208s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.152521] env[61728]: DEBUG nova.objects.instance [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lazy-loading 'resources' on Instance uuid 3c4f7d53-6ba6-4e2c-b324-e651640b304b {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 602.236204] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463684, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064106} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.236204] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 602.236656] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383a4e50-379b-44be-b8b8-214611cafdb2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.258160] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Reconfiguring VM instance instance-00000019 to attach disk [datastore1] fc3c1d93-b095-4d80-8d71-243b66d85a46/fc3c1d93-b095-4d80-8d71-243b66d85a46.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 602.258423] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7176b825-395e-4da9-bd76-b61cd2d57b6e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.277587] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 602.277587] env[61728]: value = "task-463686" [ 602.277587] env[61728]: _type = "Task" [ 602.277587] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.285754] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463686, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.410544] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-463685, 'name': ReconfigVM_Task, 'duration_secs': 0.128845} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.410544] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-121945', 'volume_id': '2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'name': 'volume-2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '37b8ca5a-e0f8-414b-a363-e56db520f027', 'attached_at': '', 'detached_at': '', 'volume_id': '2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'serial': '2e85ae80-363c-4fa7-add2-31dd95e23cfd'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 602.410913] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3bd3cab-a543-4f77-9cd4-8130a861504d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.420025] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Waiting for the task: (returnval){ [ 602.420025] env[61728]: value = "task-463687" [ 602.420025] env[61728]: _type = "Task" [ 602.420025] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.428138] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-463687, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.658091] env[61728]: DEBUG nova.compute.utils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.662056] env[61728]: DEBUG nova.compute.manager [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 602.662314] env[61728]: DEBUG nova.network.neutron [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 602.726542] env[61728]: DEBUG nova.policy [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e38d19bc24c400da6f3ef9638be0646', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b3cb7d751534c7785d758c78886e207', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 602.746739] env[61728]: DEBUG nova.network.neutron [req-6248bfbf-1fe2-4306-8e27-37d2646f157e req-fb46ada2-2f77-431a-9d1c-9c9d0978c06f service nova] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updated VIF entry in instance network info cache for port 6c9a6b61-5248-4996-a390-c18edb3ec021. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 602.747111] env[61728]: DEBUG nova.network.neutron [req-6248bfbf-1fe2-4306-8e27-37d2646f157e req-fb46ada2-2f77-431a-9d1c-9c9d0978c06f service nova] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance_info_cache with network_info: [{"id": "6c9a6b61-5248-4996-a390-c18edb3ec021", "address": "fa:16:3e:43:9a:d5", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9a6b61-52", "ovs_interfaceid": "6c9a6b61-5248-4996-a390-c18edb3ec021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.791293] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463686, 'name': ReconfigVM_Task, 'duration_secs': 0.285517} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.791543] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Reconfigured VM instance instance-00000019 to attach disk [datastore1] fc3c1d93-b095-4d80-8d71-243b66d85a46/fc3c1d93-b095-4d80-8d71-243b66d85a46.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 602.792193] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95b601ce-7325-4f7e-b992-0b80478b1831 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.799725] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 602.799725] env[61728]: value = "task-463688" [ 602.799725] env[61728]: _type = "Task" [ 602.799725] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.808184] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463688, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.930863] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-463687, 'name': Rename_Task, 'duration_secs': 0.170195} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.931250] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 602.931593] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b0b383e-5ac0-4f1b-8809-cd4942e04573 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.939319] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Waiting for the task: (returnval){ [ 602.939319] env[61728]: value = "task-463689" [ 602.939319] env[61728]: _type = "Task" [ 602.939319] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.962399] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-463689, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.136418] env[61728]: DEBUG nova.network.neutron [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Successfully created port: 502b04b6-ffe9-4337-9574-426293d93c4c {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 603.165258] env[61728]: DEBUG nova.compute.manager [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 603.234131] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374e6f64-7292-422c-886f-964da2d4fc14 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.241737] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef03be31-a9b5-43df-8fe4-0bb91bb9cae0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.250978] env[61728]: DEBUG oslo_concurrency.lockutils [req-6248bfbf-1fe2-4306-8e27-37d2646f157e req-fb46ada2-2f77-431a-9d1c-9c9d0978c06f service nova] Releasing lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.277513] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034d5051-5b00-4935-93d8-c5d56f62824f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.289735] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ed1014-a245-41ce-918a-119e9ae8b663 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.304113] env[61728]: DEBUG nova.compute.provider_tree [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.316049] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463688, 'name': Rename_Task, 'duration_secs': 0.145884} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.317063] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 603.319493] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70a2a6c4-2940-4b5a-ab64-7c8a7a825c0a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.325850] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 603.325850] env[61728]: value = "task-463690" [ 603.325850] env[61728]: _type = "Task" [ 603.325850] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.334397] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463690, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.451442] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-463689, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.808339] env[61728]: DEBUG nova.scheduler.client.report [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.837096] env[61728]: DEBUG oslo_vmware.api [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463690, 'name': PowerOnVM_Task, 'duration_secs': 0.463697} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.837096] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 603.837323] env[61728]: INFO nova.compute.manager [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Took 8.49 seconds to spawn the instance on the hypervisor. [ 603.837373] env[61728]: DEBUG nova.compute.manager [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 603.838143] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d8b38d-01bf-47ec-abc4-f5dd716a0316 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.951009] env[61728]: DEBUG oslo_vmware.api [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-463689, 'name': PowerOnVM_Task, 'duration_secs': 0.598154} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.951294] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 603.951490] env[61728]: INFO nova.compute.manager [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Took 4.21 seconds to spawn the instance on the hypervisor. [ 603.951693] env[61728]: DEBUG nova.compute.manager [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 603.952535] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21913c1-1f23-4297-85d1-f71ae38d7acf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.177356] env[61728]: DEBUG nova.compute.manager [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.204731] env[61728]: DEBUG nova.virt.hardware [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.205068] env[61728]: DEBUG nova.virt.hardware [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.205291] env[61728]: DEBUG nova.virt.hardware [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.205524] env[61728]: DEBUG nova.virt.hardware [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.205715] env[61728]: DEBUG nova.virt.hardware [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.206028] env[61728]: DEBUG nova.virt.hardware [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.206390] env[61728]: DEBUG nova.virt.hardware [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.206861] env[61728]: DEBUG nova.virt.hardware [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.207070] env[61728]: DEBUG nova.virt.hardware [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.207292] env[61728]: DEBUG nova.virt.hardware [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.207476] env[61728]: DEBUG nova.virt.hardware [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.208351] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfeb4ac8-ce23-48ae-90bd-18420d9c3ac4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.217338] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda079a8-1647-446c-b627-68851010cca6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.317385] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.162s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.317385] env[61728]: DEBUG oslo_concurrency.lockutils [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.445s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.317385] env[61728]: DEBUG nova.objects.instance [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Lazy-loading 'resources' on Instance uuid be014f77-53ca-42a5-9255-2ba950cca314 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 604.337536] env[61728]: INFO nova.scheduler.client.report [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Deleted allocations for instance 3c4f7d53-6ba6-4e2c-b324-e651640b304b [ 604.356262] env[61728]: INFO nova.compute.manager [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Took 44.09 seconds to build instance. [ 604.469164] env[61728]: INFO nova.compute.manager [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Took 43.82 seconds to build instance. [ 604.849535] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7ce73b31-8b64-45a7-9ab8-06257433c6bd tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "3c4f7d53-6ba6-4e2c-b324-e651640b304b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.359s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.858804] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a33a292-35ac-4711-8a10-909353dd34a2 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "fc3c1d93-b095-4d80-8d71-243b66d85a46" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.091s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.975331] env[61728]: DEBUG oslo_concurrency.lockutils [None req-35aa95b9-8960-49df-a9f1-40b0a4fb93cd tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Lock "37b8ca5a-e0f8-414b-a363-e56db520f027" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.444s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.094572] env[61728]: DEBUG nova.compute.manager [req-15f5aa16-07a6-4efc-a389-169e8efc0216 req-d853de84-97e6-47ac-b463-4cd608ac7910 service nova] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Received event network-vif-plugged-502b04b6-ffe9-4337-9574-426293d93c4c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 605.094795] env[61728]: DEBUG oslo_concurrency.lockutils [req-15f5aa16-07a6-4efc-a389-169e8efc0216 req-d853de84-97e6-47ac-b463-4cd608ac7910 service nova] Acquiring lock "f947d808-ebea-4593-b8e6-d3851c5216dc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.095032] env[61728]: DEBUG oslo_concurrency.lockutils [req-15f5aa16-07a6-4efc-a389-169e8efc0216 req-d853de84-97e6-47ac-b463-4cd608ac7910 service nova] Lock "f947d808-ebea-4593-b8e6-d3851c5216dc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.095407] env[61728]: DEBUG oslo_concurrency.lockutils [req-15f5aa16-07a6-4efc-a389-169e8efc0216 req-d853de84-97e6-47ac-b463-4cd608ac7910 service nova] Lock "f947d808-ebea-4593-b8e6-d3851c5216dc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.095587] env[61728]: DEBUG nova.compute.manager [req-15f5aa16-07a6-4efc-a389-169e8efc0216 req-d853de84-97e6-47ac-b463-4cd608ac7910 service nova] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] No waiting events found dispatching network-vif-plugged-502b04b6-ffe9-4337-9574-426293d93c4c {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 605.095760] env[61728]: WARNING nova.compute.manager [req-15f5aa16-07a6-4efc-a389-169e8efc0216 req-d853de84-97e6-47ac-b463-4cd608ac7910 service nova] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Received unexpected event network-vif-plugged-502b04b6-ffe9-4337-9574-426293d93c4c for instance with vm_state building and task_state spawning. [ 605.180335] env[61728]: DEBUG nova.network.neutron [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Successfully updated port: 502b04b6-ffe9-4337-9574-426293d93c4c {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 605.363345] env[61728]: DEBUG nova.compute.manager [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 605.471950] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c68aa4-7f02-4e99-8450-cb16ff3020d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.475046] env[61728]: DEBUG nova.compute.manager [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 605.482677] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7306573-924b-4d34-ae01-8d635fc463a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.514236] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dba9f88-3a62-4479-aef2-ad0584d800a3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.522049] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ed7802e-7d76-4040-9a40-3eb1cda8e8ea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.536034] env[61728]: DEBUG nova.compute.provider_tree [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.603608] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3491b532-7320-4651-9f54-810d8a857dd0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "fc3c1d93-b095-4d80-8d71-243b66d85a46" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.603932] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3491b532-7320-4651-9f54-810d8a857dd0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "fc3c1d93-b095-4d80-8d71-243b66d85a46" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.604175] env[61728]: DEBUG nova.compute.manager [None req-3491b532-7320-4651-9f54-810d8a857dd0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 605.606648] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c922c7-9b4f-485d-8373-8b7915c75953 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.614239] env[61728]: DEBUG nova.compute.manager [None req-3491b532-7320-4651-9f54-810d8a857dd0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61728) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 605.614751] env[61728]: DEBUG nova.objects.instance [None req-3491b532-7320-4651-9f54-810d8a857dd0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lazy-loading 'flavor' on Instance uuid fc3c1d93-b095-4d80-8d71-243b66d85a46 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 605.689066] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Acquiring lock "refresh_cache-f947d808-ebea-4593-b8e6-d3851c5216dc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.689066] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Acquired lock "refresh_cache-f947d808-ebea-4593-b8e6-d3851c5216dc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.689066] env[61728]: DEBUG nova.network.neutron [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 605.886399] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.995786] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.039143] env[61728]: DEBUG nova.scheduler.client.report [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.121236] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3491b532-7320-4651-9f54-810d8a857dd0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 606.121236] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b494238f-24cc-4c65-942e-6efc28f0b714 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.128588] env[61728]: DEBUG oslo_vmware.api [None req-3491b532-7320-4651-9f54-810d8a857dd0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 606.128588] env[61728]: value = "task-463691" [ 606.128588] env[61728]: _type = "Task" [ 606.128588] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.137259] env[61728]: DEBUG oslo_vmware.api [None req-3491b532-7320-4651-9f54-810d8a857dd0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463691, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.234968] env[61728]: DEBUG nova.network.neutron [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.342194] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "ca8d41dc-5719-4ade-b82a-b1d7f1c8def1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.342600] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "ca8d41dc-5719-4ade-b82a-b1d7f1c8def1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.431588] env[61728]: DEBUG nova.network.neutron [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Updating instance_info_cache with network_info: [{"id": "502b04b6-ffe9-4337-9574-426293d93c4c", "address": "fa:16:3e:9c:bf:4d", "network": {"id": "e82b56e7-d5ae-4782-8680-6730888748fe", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-269802360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b3cb7d751534c7785d758c78886e207", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap502b04b6-ff", "ovs_interfaceid": "502b04b6-ffe9-4337-9574-426293d93c4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.546102] env[61728]: DEBUG oslo_concurrency.lockutils [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.229s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.548616] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.958s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.548848] env[61728]: DEBUG nova.objects.instance [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lazy-loading 'resources' on Instance uuid e56a59a6-fc0f-45ca-810a-76b18bae173a {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 606.564952] env[61728]: INFO nova.scheduler.client.report [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Deleted allocations for instance be014f77-53ca-42a5-9255-2ba950cca314 [ 606.640011] env[61728]: DEBUG oslo_vmware.api [None req-3491b532-7320-4651-9f54-810d8a857dd0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463691, 'name': PowerOffVM_Task, 'duration_secs': 0.316757} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.640347] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3491b532-7320-4651-9f54-810d8a857dd0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 606.640545] env[61728]: DEBUG nova.compute.manager [None req-3491b532-7320-4651-9f54-810d8a857dd0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 606.641372] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4712a724-aef7-4b13-8960-5cc137005776 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.934801] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Releasing lock "refresh_cache-f947d808-ebea-4593-b8e6-d3851c5216dc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.935212] env[61728]: DEBUG nova.compute.manager [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Instance network_info: |[{"id": "502b04b6-ffe9-4337-9574-426293d93c4c", "address": "fa:16:3e:9c:bf:4d", "network": {"id": "e82b56e7-d5ae-4782-8680-6730888748fe", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-269802360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b3cb7d751534c7785d758c78886e207", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap502b04b6-ff", "ovs_interfaceid": "502b04b6-ffe9-4337-9574-426293d93c4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 606.935670] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:bf:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d0c6fd7-3cc9-4818-9475-8f15900394cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '502b04b6-ffe9-4337-9574-426293d93c4c', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 606.943888] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Creating folder: Project (7b3cb7d751534c7785d758c78886e207). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 606.944522] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6e321df1-2b1a-4acb-9d5d-a5f345db11a5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.955553] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Created folder: Project (7b3cb7d751534c7785d758c78886e207) in parent group-v121913. [ 606.955742] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Creating folder: Instances. Parent ref: group-v121996. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 606.955977] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ca659d06-f35d-491e-adbb-73cef0c4b46b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.963653] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Created folder: Instances in parent group-v121996. [ 606.963908] env[61728]: DEBUG oslo.service.loopingcall [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 606.964125] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 606.964329] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-27de7ac3-ff47-4ec5-9483-4c1f99b5354b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.982707] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 606.982707] env[61728]: value = "task-463694" [ 606.982707] env[61728]: _type = "Task" [ 606.982707] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.990622] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463694, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.074334] env[61728]: DEBUG oslo_concurrency.lockutils [None req-efa96d28-03f2-4227-bd32-b5ba1cd3b61d tempest-ServersAaction247Test-1958370266 tempest-ServersAaction247Test-1958370266-project-member] Lock "be014f77-53ca-42a5-9255-2ba950cca314" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.957s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.158578] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3491b532-7320-4651-9f54-810d8a857dd0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "fc3c1d93-b095-4d80-8d71-243b66d85a46" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.555s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.179666] env[61728]: DEBUG nova.compute.manager [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Received event network-changed-b797bbf1-77c6-48d3-b63c-f4caf481f773 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 607.179868] env[61728]: DEBUG nova.compute.manager [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Refreshing instance network info cache due to event network-changed-b797bbf1-77c6-48d3-b63c-f4caf481f773. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 607.180170] env[61728]: DEBUG oslo_concurrency.lockutils [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] Acquiring lock "refresh_cache-37b8ca5a-e0f8-414b-a363-e56db520f027" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.180403] env[61728]: DEBUG oslo_concurrency.lockutils [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] Acquired lock "refresh_cache-37b8ca5a-e0f8-414b-a363-e56db520f027" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.180750] env[61728]: DEBUG nova.network.neutron [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Refreshing network info cache for port b797bbf1-77c6-48d3-b63c-f4caf481f773 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 607.494963] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463694, 'name': CreateVM_Task, 'duration_secs': 0.426776} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.495155] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 607.495773] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.495936] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.496262] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 607.497220] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f75ad95-d905-4652-8e26-67c39e72dad2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.499159] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223d5424-4b01-491f-ae5e-5badc0d5a00f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.503916] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 607.503916] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521850fc-87a1-bc36-1bad-519f95b23901" [ 607.503916] env[61728]: _type = "Task" [ 607.503916] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.509182] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66bc0e89-9006-4f9b-a120-da0a3bb1fea4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.517264] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521850fc-87a1-bc36-1bad-519f95b23901, 'name': SearchDatastore_Task, 'duration_secs': 0.008982} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.540478] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.540676] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 607.540858] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.541016] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.541203] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 607.541638] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca1476c9-b132-44f0-b8bf-df1434a2bacc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.543915] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62d3012-216e-45fc-aed8-1017a2422e2c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.551019] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02243270-32e8-4f31-bbe5-ec067422f3c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.555511] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 607.555684] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 607.556674] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbd5d4d0-26c1-4b7f-888c-65beba90a437 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.566637] env[61728]: DEBUG nova.compute.provider_tree [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.570509] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 607.570509] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f6a48d-332c-f32e-5c01-85c8d908e9d8" [ 607.570509] env[61728]: _type = "Task" [ 607.570509] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.577384] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f6a48d-332c-f32e-5c01-85c8d908e9d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.037215] env[61728]: DEBUG nova.network.neutron [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Updated VIF entry in instance network info cache for port b797bbf1-77c6-48d3-b63c-f4caf481f773. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 608.037729] env[61728]: DEBUG nova.network.neutron [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Updating instance_info_cache with network_info: [{"id": "b797bbf1-77c6-48d3-b63c-f4caf481f773", "address": "fa:16:3e:ed:af:31", "network": {"id": "cd3b3b4d-d9ae-4037-825a-5693706df99c", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-2112414125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1aaff1dd21b497eba4d4dc9643534db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c8a5d7c-ee1f-4a41-94e4-db31e85a398d", "external-id": "cl2-zone-613", "segmentation_id": 613, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb797bbf1-77", "ovs_interfaceid": "b797bbf1-77c6-48d3-b63c-f4caf481f773", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.071447] env[61728]: DEBUG nova.scheduler.client.report [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.087219] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f6a48d-332c-f32e-5c01-85c8d908e9d8, 'name': SearchDatastore_Task, 'duration_secs': 0.009152} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.088774] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42098d75-e73a-43e1-a4f4-50205f9329a5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.099149] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 608.099149] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c989c7-fbb0-fe5d-9ce0-7f2238cf5f6d" [ 608.099149] env[61728]: _type = "Task" [ 608.099149] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.107207] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c989c7-fbb0-fe5d-9ce0-7f2238cf5f6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.548945] env[61728]: DEBUG oslo_concurrency.lockutils [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] Releasing lock "refresh_cache-37b8ca5a-e0f8-414b-a363-e56db520f027" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.549251] env[61728]: DEBUG nova.compute.manager [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Received event network-changed-502b04b6-ffe9-4337-9574-426293d93c4c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 608.549427] env[61728]: DEBUG nova.compute.manager [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Refreshing instance network info cache due to event network-changed-502b04b6-ffe9-4337-9574-426293d93c4c. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 608.549671] env[61728]: DEBUG oslo_concurrency.lockutils [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] Acquiring lock "refresh_cache-f947d808-ebea-4593-b8e6-d3851c5216dc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.549817] env[61728]: DEBUG oslo_concurrency.lockutils [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] Acquired lock "refresh_cache-f947d808-ebea-4593-b8e6-d3851c5216dc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.549978] env[61728]: DEBUG nova.network.neutron [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Refreshing network info cache for port 502b04b6-ffe9-4337-9574-426293d93c4c {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 608.580690] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.032s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.583266] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.623s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.585138] env[61728]: INFO nova.compute.claims [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.612995] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c989c7-fbb0-fe5d-9ce0-7f2238cf5f6d, 'name': SearchDatastore_Task, 'duration_secs': 0.008913} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.613122] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.613328] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] f947d808-ebea-4593-b8e6-d3851c5216dc/f947d808-ebea-4593-b8e6-d3851c5216dc.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 608.613632] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-568d5e68-ed06-4a17-b483-e1ca8f58562b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.620860] env[61728]: INFO nova.scheduler.client.report [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Deleted allocations for instance e56a59a6-fc0f-45ca-810a-76b18bae173a [ 608.621507] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 608.621507] env[61728]: value = "task-463695" [ 608.621507] env[61728]: _type = "Task" [ 608.621507] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.633729] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463695, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.079701] env[61728]: DEBUG nova.compute.manager [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 609.081096] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bdc9c02-a4f1-4ea1-9153-53d19a043c3b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.135807] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463695, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.4662} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.135918] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07d938c4-0fb9-4010-b551-54d39096916a tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "e56a59a6-fc0f-45ca-810a-76b18bae173a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.465s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.136901] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] f947d808-ebea-4593-b8e6-d3851c5216dc/f947d808-ebea-4593-b8e6-d3851c5216dc.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 609.136992] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 609.137600] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dca83839-aea2-42ae-b46e-f7c5b7b236ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.144768] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 609.144768] env[61728]: value = "task-463696" [ 609.144768] env[61728]: _type = "Task" [ 609.144768] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.156825] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463696, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.306609] env[61728]: DEBUG nova.network.neutron [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Updated VIF entry in instance network info cache for port 502b04b6-ffe9-4337-9574-426293d93c4c. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 609.306960] env[61728]: DEBUG nova.network.neutron [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Updating instance_info_cache with network_info: [{"id": "502b04b6-ffe9-4337-9574-426293d93c4c", "address": "fa:16:3e:9c:bf:4d", "network": {"id": "e82b56e7-d5ae-4782-8680-6730888748fe", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-269802360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b3cb7d751534c7785d758c78886e207", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap502b04b6-ff", "ovs_interfaceid": "502b04b6-ffe9-4337-9574-426293d93c4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.601024] env[61728]: INFO nova.compute.manager [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] instance snapshotting [ 609.601160] env[61728]: WARNING nova.compute.manager [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 609.604659] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41fd794e-5f5f-473b-ad3f-2b3a6fa9d819 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.625420] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75eb26b-1200-4fae-90f8-8c848f69b498 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.655697] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463696, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06044} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.655958] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 609.658683] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17cd30f0-47db-43b1-bb92-97cc3ad43515 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.678547] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Reconfiguring VM instance instance-0000001b to attach disk [datastore1] f947d808-ebea-4593-b8e6-d3851c5216dc/f947d808-ebea-4593-b8e6-d3851c5216dc.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 609.681932] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-105f9679-6f0f-4ebb-8928-484f02820957 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.701877] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 609.701877] env[61728]: value = "task-463697" [ 609.701877] env[61728]: _type = "Task" [ 609.701877] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.713485] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463697, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.811358] env[61728]: DEBUG oslo_concurrency.lockutils [req-3541c0f8-fad3-4561-ab38-203aa179cc30 req-8d568756-395c-4d82-bce9-44860ebb94f7 service nova] Releasing lock "refresh_cache-f947d808-ebea-4593-b8e6-d3851c5216dc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.971660] env[61728]: DEBUG oslo_concurrency.lockutils [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "adb58292-62cd-4142-ad66-db014bcbdc49" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.972072] env[61728]: DEBUG oslo_concurrency.lockutils [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "adb58292-62cd-4142-ad66-db014bcbdc49" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.972375] env[61728]: DEBUG oslo_concurrency.lockutils [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "adb58292-62cd-4142-ad66-db014bcbdc49-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.972852] env[61728]: DEBUG oslo_concurrency.lockutils [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "adb58292-62cd-4142-ad66-db014bcbdc49-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.973252] env[61728]: DEBUG oslo_concurrency.lockutils [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "adb58292-62cd-4142-ad66-db014bcbdc49-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.975534] env[61728]: INFO nova.compute.manager [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Terminating instance [ 609.981020] env[61728]: DEBUG nova.compute.manager [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 609.981020] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 609.981020] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229f0204-90a0-45d0-93d0-6b98216eccef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.991600] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 609.991687] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-04c5f4ed-432e-4074-b60a-6bcfa23f4bdc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.999700] env[61728]: DEBUG oslo_vmware.api [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 609.999700] env[61728]: value = "task-463698" [ 609.999700] env[61728]: _type = "Task" [ 609.999700] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.009007] env[61728]: DEBUG oslo_vmware.api [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463698, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.141222] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 610.144020] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-87a4c92a-6422-449f-a1d7-8961fa6af4d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.151897] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 610.151897] env[61728]: value = "task-463699" [ 610.151897] env[61728]: _type = "Task" [ 610.151897] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.165107] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463699, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.187968] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d45872a-97a1-4947-bc78-b2be589d73ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.196243] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bf2e71-7e7b-487d-a51a-c6d9d35a908b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.231034] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35371091-b087-4b1e-8faa-e0e7a1361b56 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.236329] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463697, 'name': ReconfigVM_Task, 'duration_secs': 0.318729} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.236933] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Reconfigured VM instance instance-0000001b to attach disk [datastore1] f947d808-ebea-4593-b8e6-d3851c5216dc/f947d808-ebea-4593-b8e6-d3851c5216dc.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 610.237545] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8f64736-49ac-4916-860b-757cbb366b4e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.242104] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d32393-2a1a-4b4b-9fb8-1d0231a85513 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.247060] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 610.247060] env[61728]: value = "task-463700" [ 610.247060] env[61728]: _type = "Task" [ 610.247060] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.257671] env[61728]: DEBUG nova.compute.provider_tree [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.263610] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463700, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.510169] env[61728]: DEBUG oslo_vmware.api [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463698, 'name': PowerOffVM_Task, 'duration_secs': 0.197636} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.510426] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 610.510593] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 610.510830] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4fbfa454-5906-483f-aa7e-60ca1e8c5ce2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.571671] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 610.571899] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 610.572092] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Deleting the datastore file [datastore1] adb58292-62cd-4142-ad66-db014bcbdc49 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 610.572345] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b459df70-e0dd-40f9-9365-567d6d2118d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.578588] env[61728]: DEBUG oslo_vmware.api [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 610.578588] env[61728]: value = "task-463702" [ 610.578588] env[61728]: _type = "Task" [ 610.578588] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.586254] env[61728]: DEBUG oslo_vmware.api [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463702, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.661310] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463699, 'name': CreateSnapshot_Task, 'duration_secs': 0.408439} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.661677] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 610.662503] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb24b897-22bb-4494-8e12-13dd8d756a9f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.756593] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463700, 'name': Rename_Task, 'duration_secs': 0.144747} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.756871] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 610.757358] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-567fb7b7-cbfa-4ff5-9020-5ec10a60a79e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.760840] env[61728]: DEBUG nova.scheduler.client.report [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.765655] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 610.765655] env[61728]: value = "task-463703" [ 610.765655] env[61728]: _type = "Task" [ 610.765655] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.773060] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463703, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.088385] env[61728]: DEBUG oslo_vmware.api [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463702, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132128} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.088672] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 611.088928] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 611.089076] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 611.089252] env[61728]: INFO nova.compute.manager [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Took 1.11 seconds to destroy the instance on the hypervisor. [ 611.089671] env[61728]: DEBUG oslo.service.loopingcall [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.089734] env[61728]: DEBUG nova.compute.manager [-] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.089803] env[61728]: DEBUG nova.network.neutron [-] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 611.180103] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 611.180362] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-065498a9-ac76-4b0a-ac8e-fb37e4ca7c88 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.191457] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 611.191457] env[61728]: value = "task-463704" [ 611.191457] env[61728]: _type = "Task" [ 611.191457] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.200729] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463704, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.270090] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.685s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.270090] env[61728]: DEBUG nova.compute.manager [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 611.272162] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 32.259s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.272322] env[61728]: DEBUG nova.objects.instance [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61728) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 611.298241] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463703, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.411078] env[61728]: DEBUG nova.compute.manager [req-b4cd1acd-c536-4a84-aa77-8b7a06c6dd70 req-022785e8-9a82-47e3-94ba-b7c55571076a service nova] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Received event network-vif-deleted-da4f38d4-527c-47c4-9fa7-a5009a0a4e6c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 611.411078] env[61728]: INFO nova.compute.manager [req-b4cd1acd-c536-4a84-aa77-8b7a06c6dd70 req-022785e8-9a82-47e3-94ba-b7c55571076a service nova] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Neutron deleted interface da4f38d4-527c-47c4-9fa7-a5009a0a4e6c; detaching it from the instance and deleting it from the info cache [ 611.411078] env[61728]: DEBUG nova.network.neutron [req-b4cd1acd-c536-4a84-aa77-8b7a06c6dd70 req-022785e8-9a82-47e3-94ba-b7c55571076a service nova] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.701847] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463704, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.780561] env[61728]: DEBUG oslo_vmware.api [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463703, 'name': PowerOnVM_Task, 'duration_secs': 0.783051} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.780862] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 611.781086] env[61728]: INFO nova.compute.manager [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Took 7.60 seconds to spawn the instance on the hypervisor. [ 611.781267] env[61728]: DEBUG nova.compute.manager [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 611.782075] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76608cf0-96d4-4cc3-943e-fd4b6f22b559 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.785893] env[61728]: DEBUG nova.compute.utils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 611.787266] env[61728]: DEBUG nova.compute.manager [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 611.787493] env[61728]: DEBUG nova.network.neutron [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 611.845365] env[61728]: DEBUG nova.policy [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8a56b87b2294bc8b71a599c076414d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8fc16f1d1b144286b180cd7c8f7588fd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 611.885961] env[61728]: DEBUG nova.network.neutron [-] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.916024] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-408bfa02-d567-43cb-8da7-0e33bc39cafe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.928697] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2362c898-c696-4dd4-802c-4be328d4de54 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.964312] env[61728]: DEBUG nova.compute.manager [req-b4cd1acd-c536-4a84-aa77-8b7a06c6dd70 req-022785e8-9a82-47e3-94ba-b7c55571076a service nova] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Detach interface failed, port_id=da4f38d4-527c-47c4-9fa7-a5009a0a4e6c, reason: Instance adb58292-62cd-4142-ad66-db014bcbdc49 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 612.204303] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463704, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.293428] env[61728]: DEBUG nova.compute.manager [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 612.296987] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f1139d90-4489-42f6-bde0-3642ee659982 tempest-ServersAdmin275Test-1630563156 tempest-ServersAdmin275Test-1630563156-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.025s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.298199] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.643s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.302312] env[61728]: INFO nova.compute.claims [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.308052] env[61728]: DEBUG nova.network.neutron [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Successfully created port: ccffc76c-1df1-4ba7-824a-5c175febeca3 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 612.322524] env[61728]: INFO nova.compute.manager [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Took 43.25 seconds to build instance. [ 612.391673] env[61728]: INFO nova.compute.manager [-] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Took 1.30 seconds to deallocate network for instance. [ 612.703443] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463704, 'name': CloneVM_Task, 'duration_secs': 1.404711} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.703700] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Created linked-clone VM from snapshot [ 612.704436] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b24a4f8-d355-42eb-b414-1a0b46e911a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.711450] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Uploading image 4b0aeaca-e578-41be-bb8c-6d3ee114d452 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 612.732485] env[61728]: DEBUG oslo_vmware.rw_handles [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 612.732485] env[61728]: value = "vm-122000" [ 612.732485] env[61728]: _type = "VirtualMachine" [ 612.732485] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 612.732802] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e474aa80-52f3-4c1b-9a0d-89f18613c057 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.738643] env[61728]: DEBUG oslo_vmware.rw_handles [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lease: (returnval){ [ 612.738643] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5240e0bd-c661-439b-5667-e90bbc46081d" [ 612.738643] env[61728]: _type = "HttpNfcLease" [ 612.738643] env[61728]: } obtained for exporting VM: (result){ [ 612.738643] env[61728]: value = "vm-122000" [ 612.738643] env[61728]: _type = "VirtualMachine" [ 612.738643] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 612.738865] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the lease: (returnval){ [ 612.738865] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5240e0bd-c661-439b-5667-e90bbc46081d" [ 612.738865] env[61728]: _type = "HttpNfcLease" [ 612.738865] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 612.744996] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 612.744996] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5240e0bd-c661-439b-5667-e90bbc46081d" [ 612.744996] env[61728]: _type = "HttpNfcLease" [ 612.744996] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 612.824756] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2d5c454d-3807-442f-a74f-43e6dc747761 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lock "f947d808-ebea-4593-b8e6-d3851c5216dc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.891s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.896621] env[61728]: DEBUG oslo_concurrency.lockutils [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.251720] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 613.251720] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5240e0bd-c661-439b-5667-e90bbc46081d" [ 613.251720] env[61728]: _type = "HttpNfcLease" [ 613.251720] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 613.253428] env[61728]: DEBUG oslo_vmware.rw_handles [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 613.253428] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5240e0bd-c661-439b-5667-e90bbc46081d" [ 613.253428] env[61728]: _type = "HttpNfcLease" [ 613.253428] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 613.254717] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e4fb06b-f6c3-4ec5-b305-05d70fe60d10 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.263894] env[61728]: DEBUG oslo_vmware.rw_handles [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52df4f08-9f18-c3e8-3b2b-3eb0b437a21e/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 613.264096] env[61728]: DEBUG oslo_vmware.rw_handles [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52df4f08-9f18-c3e8-3b2b-3eb0b437a21e/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 613.320525] env[61728]: DEBUG nova.compute.manager [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 613.326588] env[61728]: DEBUG nova.compute.manager [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 613.358435] env[61728]: DEBUG nova.virt.hardware [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 613.358703] env[61728]: DEBUG nova.virt.hardware [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 613.358899] env[61728]: DEBUG nova.virt.hardware [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 613.359145] env[61728]: DEBUG nova.virt.hardware [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 613.359313] env[61728]: DEBUG nova.virt.hardware [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 613.359463] env[61728]: DEBUG nova.virt.hardware [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 613.359686] env[61728]: DEBUG nova.virt.hardware [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 613.359860] env[61728]: DEBUG nova.virt.hardware [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 613.360134] env[61728]: DEBUG nova.virt.hardware [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 613.360198] env[61728]: DEBUG nova.virt.hardware [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 613.360378] env[61728]: DEBUG nova.virt.hardware [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 613.360891] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-353d4ad7-f52d-4c51-a2cf-ae9407ef3773 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.363432] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fedb790c-2d60-428a-8231-95ec37643fc5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.376963] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b730f114-094f-471b-bbb0-fde1575a37f9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.795919] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9015469-a77c-4df0-85c8-ce22c3317551 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.804111] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6c7185-53bc-406a-a239-4fcb70b575c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.850861] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514c8a90-2fc5-4360-8227-b750487738d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.864742] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d34191-d3dc-4039-bc4d-ccc08700a0fa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.882748] env[61728]: DEBUG nova.compute.provider_tree [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.885772] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.994059] env[61728]: DEBUG nova.compute.manager [req-960dfd6b-2fa6-4194-a070-8633c3374962 req-f285918d-73b4-4f97-a5e9-aee64d000450 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Received event network-vif-plugged-ccffc76c-1df1-4ba7-824a-5c175febeca3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 613.994364] env[61728]: DEBUG oslo_concurrency.lockutils [req-960dfd6b-2fa6-4194-a070-8633c3374962 req-f285918d-73b4-4f97-a5e9-aee64d000450 service nova] Acquiring lock "c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.995918] env[61728]: DEBUG oslo_concurrency.lockutils [req-960dfd6b-2fa6-4194-a070-8633c3374962 req-f285918d-73b4-4f97-a5e9-aee64d000450 service nova] Lock "c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.996231] env[61728]: DEBUG oslo_concurrency.lockutils [req-960dfd6b-2fa6-4194-a070-8633c3374962 req-f285918d-73b4-4f97-a5e9-aee64d000450 service nova] Lock "c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.996372] env[61728]: DEBUG nova.compute.manager [req-960dfd6b-2fa6-4194-a070-8633c3374962 req-f285918d-73b4-4f97-a5e9-aee64d000450 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] No waiting events found dispatching network-vif-plugged-ccffc76c-1df1-4ba7-824a-5c175febeca3 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 613.996602] env[61728]: WARNING nova.compute.manager [req-960dfd6b-2fa6-4194-a070-8633c3374962 req-f285918d-73b4-4f97-a5e9-aee64d000450 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Received unexpected event network-vif-plugged-ccffc76c-1df1-4ba7-824a-5c175febeca3 for instance with vm_state building and task_state spawning. [ 614.073643] env[61728]: DEBUG nova.network.neutron [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Successfully updated port: ccffc76c-1df1-4ba7-824a-5c175febeca3 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 614.389435] env[61728]: DEBUG nova.scheduler.client.report [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.576989] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "refresh_cache-c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.577163] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquired lock "refresh_cache-c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.577324] env[61728]: DEBUG nova.network.neutron [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 614.861171] env[61728]: DEBUG nova.compute.manager [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 614.862191] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-191cde25-2f02-4737-b818-6235f5e34cb3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.896900] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.599s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.896900] env[61728]: DEBUG nova.compute.manager [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 614.900962] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.462s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.901386] env[61728]: DEBUG nova.objects.instance [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Lazy-loading 'resources' on Instance uuid a0831461-ece1-43ee-92f6-34d7d4e673e2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 615.114357] env[61728]: DEBUG nova.network.neutron [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.318610] env[61728]: DEBUG nova.network.neutron [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Updating instance_info_cache with network_info: [{"id": "ccffc76c-1df1-4ba7-824a-5c175febeca3", "address": "fa:16:3e:33:14:5f", "network": {"id": "5d1e643d-f8a4-4302-9920-bb661df4632d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1891316321-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fc16f1d1b144286b180cd7c8f7588fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapccffc76c-1d", "ovs_interfaceid": "ccffc76c-1df1-4ba7-824a-5c175febeca3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.373862] env[61728]: INFO nova.compute.manager [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] instance snapshotting [ 615.376870] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a05abe-b354-465d-9230-d408c7783e3d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.396454] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3262f137-139f-436e-9037-d857d0b1bad4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.405126] env[61728]: DEBUG nova.compute.utils [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.410620] env[61728]: DEBUG nova.compute.manager [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 615.410812] env[61728]: DEBUG nova.network.neutron [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 615.469223] env[61728]: DEBUG nova.policy [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1217ab78b9334f949ef0430ab1484efb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '667c26f99bea48cabb7d0bed8fa13b1a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 615.821354] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Releasing lock "refresh_cache-c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.821688] env[61728]: DEBUG nova.compute.manager [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Instance network_info: |[{"id": "ccffc76c-1df1-4ba7-824a-5c175febeca3", "address": "fa:16:3e:33:14:5f", "network": {"id": "5d1e643d-f8a4-4302-9920-bb661df4632d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1891316321-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fc16f1d1b144286b180cd7c8f7588fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapccffc76c-1d", "ovs_interfaceid": "ccffc76c-1df1-4ba7-824a-5c175febeca3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 615.822118] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:14:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '877ed63d-906e-4bd5-a1fc-7e82d172d41e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ccffc76c-1df1-4ba7-824a-5c175febeca3', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 615.834733] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Creating folder: Project (8fc16f1d1b144286b180cd7c8f7588fd). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 615.837424] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6e823d01-dd3b-4549-8bfd-f23781d8005b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.848338] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Created folder: Project (8fc16f1d1b144286b180cd7c8f7588fd) in parent group-v121913. [ 615.848551] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Creating folder: Instances. Parent ref: group-v122001. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 615.851256] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fe615e59-6460-4db8-b0e5-3424586a343b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.859873] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Created folder: Instances in parent group-v122001. [ 615.862017] env[61728]: DEBUG oslo.service.loopingcall [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 615.862017] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 615.862017] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2395b6f2-3df5-42a8-9260-1e293e267e7e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.886026] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 615.886026] env[61728]: value = "task-463708" [ 615.886026] env[61728]: _type = "Task" [ 615.886026] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.895422] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463708, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.907169] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "db8e373d-e06c-43f0-a79c-2c1aaf65a01a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.907596] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "db8e373d-e06c-43f0-a79c-2c1aaf65a01a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.907947] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "db8e373d-e06c-43f0-a79c-2c1aaf65a01a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.908274] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "db8e373d-e06c-43f0-a79c-2c1aaf65a01a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.908581] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "db8e373d-e06c-43f0-a79c-2c1aaf65a01a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.910893] env[61728]: INFO nova.compute.manager [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Terminating instance [ 615.917019] env[61728]: DEBUG nova.compute.manager [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 615.917019] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 615.917019] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29cdeefa-8b55-41fd-9374-1520a8acbf5b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.918037] env[61728]: DEBUG nova.compute.manager [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 615.921636] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 615.924904] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4a0171bb-d8a6-4269-801e-52a498de4b40 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.934055] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 615.934055] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c86351de-b48d-4281-8742-ffbbefd12681 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.936235] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 615.936235] env[61728]: value = "task-463709" [ 615.936235] env[61728]: _type = "Task" [ 615.936235] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.945583] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa9d4808-dfbf-47cb-9627-03058488c778 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.953083] env[61728]: DEBUG oslo_vmware.api [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 615.953083] env[61728]: value = "task-463710" [ 615.953083] env[61728]: _type = "Task" [ 615.953083] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.962108] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463709, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.963887] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c3aa42-ffba-4814-a757-eb9aec1e1ae0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.973754] env[61728]: DEBUG oslo_vmware.api [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463710, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.003391] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a51aa29-77c4-4ffb-b44b-4c7110f1bdc6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.011402] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a9e117-55a7-40c0-86ac-7c32a9764633 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.027746] env[61728]: DEBUG nova.compute.provider_tree [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.047896] env[61728]: DEBUG nova.network.neutron [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Successfully created port: 1915184e-ab7a-409c-a8d2-ccbdca5e942e {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 616.204181] env[61728]: DEBUG nova.compute.manager [req-16a4998d-8b77-411c-91d2-b38271ba4fd1 req-781952dc-77ab-48e2-8d57-e61f0137a385 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Received event network-changed-ccffc76c-1df1-4ba7-824a-5c175febeca3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 616.204426] env[61728]: DEBUG nova.compute.manager [req-16a4998d-8b77-411c-91d2-b38271ba4fd1 req-781952dc-77ab-48e2-8d57-e61f0137a385 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Refreshing instance network info cache due to event network-changed-ccffc76c-1df1-4ba7-824a-5c175febeca3. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 616.204690] env[61728]: DEBUG oslo_concurrency.lockutils [req-16a4998d-8b77-411c-91d2-b38271ba4fd1 req-781952dc-77ab-48e2-8d57-e61f0137a385 service nova] Acquiring lock "refresh_cache-c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.205806] env[61728]: DEBUG oslo_concurrency.lockutils [req-16a4998d-8b77-411c-91d2-b38271ba4fd1 req-781952dc-77ab-48e2-8d57-e61f0137a385 service nova] Acquired lock "refresh_cache-c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.205806] env[61728]: DEBUG nova.network.neutron [req-16a4998d-8b77-411c-91d2-b38271ba4fd1 req-781952dc-77ab-48e2-8d57-e61f0137a385 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Refreshing network info cache for port ccffc76c-1df1-4ba7-824a-5c175febeca3 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 616.394662] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463708, 'name': CreateVM_Task, 'duration_secs': 0.43386} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.394845] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 616.395593] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.395785] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.396134] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 616.396409] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6951be50-cf16-48e5-b8a1-2a2dbfff591d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.401338] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 616.401338] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52945344-d614-71b9-1b14-c0f1b835a9ad" [ 616.401338] env[61728]: _type = "Task" [ 616.401338] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.411898] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52945344-d614-71b9-1b14-c0f1b835a9ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.425165] env[61728]: INFO nova.virt.block_device [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Booting with volume 2eb2f0fc-4130-44f7-bddb-eb01dbadccc4 at /dev/sda [ 616.452239] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463709, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.460590] env[61728]: DEBUG oslo_vmware.api [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463710, 'name': PowerOffVM_Task, 'duration_secs': 0.229186} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.460840] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 616.462372] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 616.462372] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4a676df6-c231-4227-986d-5f9e125e6b0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.478363] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c95ac00f-431c-4778-a9dc-df3d6ff5f15a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.487488] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b89790d-8766-49fc-b4b8-3582225f02d8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.526992] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f25d4e4b-2dea-488e-8b6b-c299df9ab10b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.529096] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 616.533123] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 616.533123] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Deleting the datastore file [datastore1] db8e373d-e06c-43f0-a79c-2c1aaf65a01a {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 616.533123] env[61728]: DEBUG nova.scheduler.client.report [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.534720] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fcc4097-0a02-4730-a2f9-d0af44735d9f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.544732] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef17926-f8f8-44e9-8feb-bba843833c6c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.557566] env[61728]: DEBUG oslo_vmware.api [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 616.557566] env[61728]: value = "task-463712" [ 616.557566] env[61728]: _type = "Task" [ 616.557566] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.568579] env[61728]: DEBUG oslo_vmware.api [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463712, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.583027] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3825c14b-e680-4bcc-92a8-b74b70874bd4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.588478] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d0d159-69ed-4aa3-addf-dc43bd0ee1ef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.605091] env[61728]: DEBUG nova.virt.block_device [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Updating existing volume attachment record: 953d8191-da35-4335-b06c-c3c847f084f1 {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 616.740729] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "b563e521-0e0e-45ea-bd5a-056285ea35b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.741076] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "b563e521-0e0e-45ea-bd5a-056285ea35b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.911993] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52945344-d614-71b9-1b14-c0f1b835a9ad, 'name': SearchDatastore_Task, 'duration_secs': 0.013425} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.912559] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.912793] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 616.913190] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.913375] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.913566] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 616.913924] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4958f87-f03f-4599-8d97-2fbaa801104d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.925157] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 616.925157] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 616.925157] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9cd2a6d9-7a7b-4b02-b085-0fd497c43602 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.929508] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 616.929508] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521131f3-d36b-56e0-f932-c1404689daa1" [ 616.929508] env[61728]: _type = "Task" [ 616.929508] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.930383] env[61728]: DEBUG nova.network.neutron [req-16a4998d-8b77-411c-91d2-b38271ba4fd1 req-781952dc-77ab-48e2-8d57-e61f0137a385 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Updated VIF entry in instance network info cache for port ccffc76c-1df1-4ba7-824a-5c175febeca3. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 616.930827] env[61728]: DEBUG nova.network.neutron [req-16a4998d-8b77-411c-91d2-b38271ba4fd1 req-781952dc-77ab-48e2-8d57-e61f0137a385 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Updating instance_info_cache with network_info: [{"id": "ccffc76c-1df1-4ba7-824a-5c175febeca3", "address": "fa:16:3e:33:14:5f", "network": {"id": "5d1e643d-f8a4-4302-9920-bb661df4632d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1891316321-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fc16f1d1b144286b180cd7c8f7588fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapccffc76c-1d", "ovs_interfaceid": "ccffc76c-1df1-4ba7-824a-5c175febeca3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.941072] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521131f3-d36b-56e0-f932-c1404689daa1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.951684] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463709, 'name': CreateSnapshot_Task, 'duration_secs': 0.729098} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.952019] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 616.953097] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a7ec9d-975c-4b88-9b3e-5be8fd34e166 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.038768] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.138s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.042248] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.051s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.043653] env[61728]: INFO nova.compute.claims [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 617.066168] env[61728]: INFO nova.scheduler.client.report [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Deleted allocations for instance a0831461-ece1-43ee-92f6-34d7d4e673e2 [ 617.070116] env[61728]: DEBUG oslo_vmware.api [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463712, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.222104} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.072223] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 617.072624] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 617.072624] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 617.073312] env[61728]: INFO nova.compute.manager [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Took 1.16 seconds to destroy the instance on the hypervisor. [ 617.073312] env[61728]: DEBUG oslo.service.loopingcall [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 617.073441] env[61728]: DEBUG nova.compute.manager [-] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 617.073980] env[61728]: DEBUG nova.network.neutron [-] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 617.437731] env[61728]: DEBUG oslo_concurrency.lockutils [req-16a4998d-8b77-411c-91d2-b38271ba4fd1 req-781952dc-77ab-48e2-8d57-e61f0137a385 service nova] Releasing lock "refresh_cache-c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.447978] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521131f3-d36b-56e0-f932-c1404689daa1, 'name': SearchDatastore_Task, 'duration_secs': 0.020333} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.448921] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5836431e-596d-4d86-be71-d2b2ebd32851 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.455277] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 617.455277] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52154875-402c-2173-9475-d4f7fbeace67" [ 617.455277] env[61728]: _type = "Task" [ 617.455277] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.464861] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52154875-402c-2173-9475-d4f7fbeace67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.473852] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 617.474227] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b252b3cc-2430-42c7-886d-0848f1b0ceee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.483355] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 617.483355] env[61728]: value = "task-463713" [ 617.483355] env[61728]: _type = "Task" [ 617.483355] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.492749] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463713, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.576689] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8c1198e3-c33a-4dd4-98e5-a96197b95d05 tempest-ServersAdmin275Test-781292626 tempest-ServersAdmin275Test-781292626-project-member] Lock "a0831461-ece1-43ee-92f6-34d7d4e673e2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.849s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.609770] env[61728]: DEBUG nova.compute.manager [req-d48a979d-6f2f-479c-a999-f408a056ed1f req-2d9bdc56-2f2c-49cf-a259-59c5138703d4 service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Received event network-vif-deleted-dd415c0f-ba4e-4019-9b70-d2788151f446 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 617.609770] env[61728]: INFO nova.compute.manager [req-d48a979d-6f2f-479c-a999-f408a056ed1f req-2d9bdc56-2f2c-49cf-a259-59c5138703d4 service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Neutron deleted interface dd415c0f-ba4e-4019-9b70-d2788151f446; detaching it from the instance and deleting it from the info cache [ 617.609770] env[61728]: DEBUG nova.network.neutron [req-d48a979d-6f2f-479c-a999-f408a056ed1f req-2d9bdc56-2f2c-49cf-a259-59c5138703d4 service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.968810] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52154875-402c-2173-9475-d4f7fbeace67, 'name': SearchDatastore_Task, 'duration_secs': 0.020715} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.969261] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.969586] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df/c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 617.969943] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-52d1bf6e-a576-4995-acb3-f9223830d17b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.977990] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 617.977990] env[61728]: value = "task-463714" [ 617.977990] env[61728]: _type = "Task" [ 617.977990] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.993314] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463714, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.999404] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463713, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.081390] env[61728]: DEBUG nova.network.neutron [-] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.111500] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dec58cd3-000c-46fc-ad65-1106170d4fcc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.122792] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5910ce4e-baa2-4d80-820d-fb3987448ca3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.162963] env[61728]: DEBUG nova.compute.manager [req-d48a979d-6f2f-479c-a999-f408a056ed1f req-2d9bdc56-2f2c-49cf-a259-59c5138703d4 service nova] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Detach interface failed, port_id=dd415c0f-ba4e-4019-9b70-d2788151f446, reason: Instance db8e373d-e06c-43f0-a79c-2c1aaf65a01a could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 618.254024] env[61728]: DEBUG nova.network.neutron [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Successfully updated port: 1915184e-ab7a-409c-a8d2-ccbdca5e942e {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 618.294392] env[61728]: DEBUG nova.compute.manager [req-feb73587-1aa9-4210-a8c8-d24f423a52af req-d2c27fab-ba42-4f46-bd8a-807c4b869f22 service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Received event network-vif-plugged-1915184e-ab7a-409c-a8d2-ccbdca5e942e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 618.294631] env[61728]: DEBUG oslo_concurrency.lockutils [req-feb73587-1aa9-4210-a8c8-d24f423a52af req-d2c27fab-ba42-4f46-bd8a-807c4b869f22 service nova] Acquiring lock "0f75629b-bdec-4593-aceb-cd0478972ccc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.295220] env[61728]: DEBUG oslo_concurrency.lockutils [req-feb73587-1aa9-4210-a8c8-d24f423a52af req-d2c27fab-ba42-4f46-bd8a-807c4b869f22 service nova] Lock "0f75629b-bdec-4593-aceb-cd0478972ccc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.295220] env[61728]: DEBUG oslo_concurrency.lockutils [req-feb73587-1aa9-4210-a8c8-d24f423a52af req-d2c27fab-ba42-4f46-bd8a-807c4b869f22 service nova] Lock "0f75629b-bdec-4593-aceb-cd0478972ccc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.295220] env[61728]: DEBUG nova.compute.manager [req-feb73587-1aa9-4210-a8c8-d24f423a52af req-d2c27fab-ba42-4f46-bd8a-807c4b869f22 service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] No waiting events found dispatching network-vif-plugged-1915184e-ab7a-409c-a8d2-ccbdca5e942e {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 618.295426] env[61728]: WARNING nova.compute.manager [req-feb73587-1aa9-4210-a8c8-d24f423a52af req-d2c27fab-ba42-4f46-bd8a-807c4b869f22 service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Received unexpected event network-vif-plugged-1915184e-ab7a-409c-a8d2-ccbdca5e942e for instance with vm_state building and task_state block_device_mapping. [ 618.498607] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463713, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.498930] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463714, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.585463] env[61728]: INFO nova.compute.manager [-] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Took 1.51 seconds to deallocate network for instance. [ 618.685716] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-094781fb-da5b-4da3-9a66-e476eec49216 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.694613] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8035aca-a7eb-444b-b93e-37b5c7b43b5b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.730270] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c3def0-159a-49b2-8dc8-7b7c8e2067c5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.740803] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e87005-69c6-4fb4-9760-90701e2709bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.748515] env[61728]: DEBUG nova.compute.manager [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 618.749358] env[61728]: DEBUG nova.virt.hardware [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.749572] env[61728]: DEBUG nova.virt.hardware [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.749572] env[61728]: DEBUG nova.virt.hardware [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.749853] env[61728]: DEBUG nova.virt.hardware [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.749971] env[61728]: DEBUG nova.virt.hardware [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.750167] env[61728]: DEBUG nova.virt.hardware [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.751103] env[61728]: DEBUG nova.virt.hardware [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.751103] env[61728]: DEBUG nova.virt.hardware [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.751103] env[61728]: DEBUG nova.virt.hardware [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.751103] env[61728]: DEBUG nova.virt.hardware [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.751349] env[61728]: DEBUG nova.virt.hardware [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.752154] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d0704a0-8c0c-4e1c-a18d-66711cc98fea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.756262] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Acquiring lock "refresh_cache-0f75629b-bdec-4593-aceb-cd0478972ccc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.756425] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Acquired lock "refresh_cache-0f75629b-bdec-4593-aceb-cd0478972ccc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.756604] env[61728]: DEBUG nova.network.neutron [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.772057] env[61728]: DEBUG nova.compute.provider_tree [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.780046] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027873c4-2d5d-4c42-9935-92a16bead53d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.817645] env[61728]: DEBUG nova.network.neutron [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.974350] env[61728]: DEBUG nova.network.neutron [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Updating instance_info_cache with network_info: [{"id": "1915184e-ab7a-409c-a8d2-ccbdca5e942e", "address": "fa:16:3e:c4:37:f4", "network": {"id": "104934bf-5809-4ef5-adb6-970f5c45ea8e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1985848519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "667c26f99bea48cabb7d0bed8fa13b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7869cc8e-e58f-4fd6-88d7-85a18e43cd3a", "external-id": "nsx-vlan-transportzone-927", "segmentation_id": 927, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1915184e-ab", "ovs_interfaceid": "1915184e-ab7a-409c-a8d2-ccbdca5e942e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.993091] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463714, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681148} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.993931] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df/c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 618.994239] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 618.994845] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-107a22af-2d9c-43b4-9d24-3401a9845ee3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.000610] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463713, 'name': CloneVM_Task} progress is 95%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.006577] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 619.006577] env[61728]: value = "task-463715" [ 619.006577] env[61728]: _type = "Task" [ 619.006577] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.015764] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463715, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.092682] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.288142] env[61728]: DEBUG nova.scheduler.client.report [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 619.477080] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Releasing lock "refresh_cache-0f75629b-bdec-4593-aceb-cd0478972ccc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.477440] env[61728]: DEBUG nova.compute.manager [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Instance network_info: |[{"id": "1915184e-ab7a-409c-a8d2-ccbdca5e942e", "address": "fa:16:3e:c4:37:f4", "network": {"id": "104934bf-5809-4ef5-adb6-970f5c45ea8e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1985848519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "667c26f99bea48cabb7d0bed8fa13b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7869cc8e-e58f-4fd6-88d7-85a18e43cd3a", "external-id": "nsx-vlan-transportzone-927", "segmentation_id": 927, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1915184e-ab", "ovs_interfaceid": "1915184e-ab7a-409c-a8d2-ccbdca5e942e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 619.477877] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:37:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7869cc8e-e58f-4fd6-88d7-85a18e43cd3a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1915184e-ab7a-409c-a8d2-ccbdca5e942e', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 619.486013] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Creating folder: Project (667c26f99bea48cabb7d0bed8fa13b1a). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 619.486427] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-378f2c01-4810-4e0d-90e6-40951d135d0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.498070] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463713, 'name': CloneVM_Task, 'duration_secs': 1.786983} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.498649] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Created linked-clone VM from snapshot [ 619.499443] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a291b6-843e-4197-b706-640a23b9c592 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.503566] env[61728]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 619.503566] env[61728]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61728) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 619.503566] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Folder already exists: Project (667c26f99bea48cabb7d0bed8fa13b1a). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 619.503741] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Creating folder: Instances. Parent ref: group-v121936. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 619.504263] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1d903029-467f-41a6-a59b-b68c4df9bf6b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.511794] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Uploading image a203be0f-1f36-4177-9319-6056c97931b5 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 619.520132] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463715, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068244} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.521676] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 619.521965] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Created folder: Instances in parent group-v121936. [ 619.522255] env[61728]: DEBUG oslo.service.loopingcall [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 619.523553] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740bce2e-7bbb-40b1-9561-5e364a92b1b6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.526272] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 619.528562] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4aad1266-113e-45d9-9857-ecfb097a2d47 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.562963] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df/c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 619.566288] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 619.566288] env[61728]: value = "vm-122005" [ 619.566288] env[61728]: _type = "VirtualMachine" [ 619.566288] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 619.566548] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e141a56a-bcfa-4e10-b9b9-38a103abebd0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.581371] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 619.581371] env[61728]: value = "task-463718" [ 619.581371] env[61728]: _type = "Task" [ 619.581371] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.581626] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7a543dd0-95c7-447a-92e1-e18aab92dccb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.588786] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 619.588786] env[61728]: value = "task-463719" [ 619.588786] env[61728]: _type = "Task" [ 619.588786] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.595931] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lease: (returnval){ [ 619.595931] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52210b98-bc41-3472-53fc-710180c7c998" [ 619.595931] env[61728]: _type = "HttpNfcLease" [ 619.595931] env[61728]: } obtained for exporting VM: (result){ [ 619.595931] env[61728]: value = "vm-122005" [ 619.595931] env[61728]: _type = "VirtualMachine" [ 619.595931] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 619.596471] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the lease: (returnval){ [ 619.596471] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52210b98-bc41-3472-53fc-710180c7c998" [ 619.596471] env[61728]: _type = "HttpNfcLease" [ 619.596471] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 619.596471] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463718, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.603078] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463719, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.606714] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 619.606714] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52210b98-bc41-3472-53fc-710180c7c998" [ 619.606714] env[61728]: _type = "HttpNfcLease" [ 619.606714] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 619.793725] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.751s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.796587] env[61728]: DEBUG nova.compute.manager [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 619.799738] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.908s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.800778] env[61728]: INFO nova.compute.claims [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 620.094056] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463718, 'name': CreateVM_Task, 'duration_secs': 0.349898} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.096641] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 620.097385] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': '953d8191-da35-4335-b06c-c3c847f084f1', 'delete_on_termination': True, 'guest_format': None, 'disk_bus': None, 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-121946', 'volume_id': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'name': 'volume-2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0f75629b-bdec-4593-aceb-cd0478972ccc', 'attached_at': '', 'detached_at': '', 'volume_id': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'serial': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4'}, 'mount_device': '/dev/sda', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=61728) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 620.097610] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Root volume attach. Driver type: vmdk {{(pid=61728) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 620.098799] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc2ead6-a627-4026-836d-2ebfcb7d2410 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.107310] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463719, 'name': ReconfigVM_Task, 'duration_secs': 0.446928} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.107465] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Reconfigured VM instance instance-0000001c to attach disk [datastore1] c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df/c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 620.108495] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ddc9fcf-9945-43a3-96b1-0f4b559f1600 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.114120] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 620.114120] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52210b98-bc41-3472-53fc-710180c7c998" [ 620.114120] env[61728]: _type = "HttpNfcLease" [ 620.114120] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 620.114784] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 620.114784] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52210b98-bc41-3472-53fc-710180c7c998" [ 620.114784] env[61728]: _type = "HttpNfcLease" [ 620.114784] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 620.115545] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6960a0b3-9855-41ea-a718-6debe8d99b72 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.118263] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36e9fdf0-240c-415c-aa9e-08f995cc82da {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.122292] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 620.122292] env[61728]: value = "task-463721" [ 620.122292] env[61728]: _type = "Task" [ 620.122292] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.129854] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c53eb3-6ff0-8632-fd43-28918162934d/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 620.130046] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c53eb3-6ff0-8632-fd43-28918162934d/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 620.132280] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e729a98d-ff2a-465e-b7a9-d02205ce08e1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.139469] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463721, 'name': Rename_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.203914] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-5b55caa0-6920-4b57-a8c3-542891767b25 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.210379] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Waiting for the task: (returnval){ [ 620.210379] env[61728]: value = "task-463722" [ 620.210379] env[61728]: _type = "Task" [ 620.210379] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.222740] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-463722, 'name': RelocateVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.236471] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2ea7822e-30a3-48e4-a8bc-ba7a8cb8e9da {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.305044] env[61728]: DEBUG nova.compute.utils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 620.308729] env[61728]: DEBUG nova.compute.manager [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 620.308925] env[61728]: DEBUG nova.network.neutron [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 620.381353] env[61728]: DEBUG nova.compute.manager [req-a051b3c2-bf3f-4be8-92dc-1773711c9de8 req-ef95b524-1cd6-4983-bc68-b17e6485a0c8 service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Received event network-changed-1915184e-ab7a-409c-a8d2-ccbdca5e942e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 620.381770] env[61728]: DEBUG nova.compute.manager [req-a051b3c2-bf3f-4be8-92dc-1773711c9de8 req-ef95b524-1cd6-4983-bc68-b17e6485a0c8 service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Refreshing instance network info cache due to event network-changed-1915184e-ab7a-409c-a8d2-ccbdca5e942e. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 620.382057] env[61728]: DEBUG oslo_concurrency.lockutils [req-a051b3c2-bf3f-4be8-92dc-1773711c9de8 req-ef95b524-1cd6-4983-bc68-b17e6485a0c8 service nova] Acquiring lock "refresh_cache-0f75629b-bdec-4593-aceb-cd0478972ccc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.382224] env[61728]: DEBUG oslo_concurrency.lockutils [req-a051b3c2-bf3f-4be8-92dc-1773711c9de8 req-ef95b524-1cd6-4983-bc68-b17e6485a0c8 service nova] Acquired lock "refresh_cache-0f75629b-bdec-4593-aceb-cd0478972ccc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.382470] env[61728]: DEBUG nova.network.neutron [req-a051b3c2-bf3f-4be8-92dc-1773711c9de8 req-ef95b524-1cd6-4983-bc68-b17e6485a0c8 service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Refreshing network info cache for port 1915184e-ab7a-409c-a8d2-ccbdca5e942e {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 620.402403] env[61728]: DEBUG nova.policy [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'df758e65d4a54121826cb6d7f00204cd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '99db3c843e8c4c3397cabba65ae510ca', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 620.643772] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463721, 'name': Rename_Task, 'duration_secs': 0.176334} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.643772] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 620.643772] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c7932cf-6528-4853-aca5-156c545db92a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.648247] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 620.648247] env[61728]: value = "task-463723" [ 620.648247] env[61728]: _type = "Task" [ 620.648247] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.667371] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463723, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.723832] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-463722, 'name': RelocateVM_Task, 'duration_secs': 0.383363} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.724474] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 620.725077] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-121946', 'volume_id': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'name': 'volume-2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0f75629b-bdec-4593-aceb-cd0478972ccc', 'attached_at': '', 'detached_at': '', 'volume_id': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'serial': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 620.726890] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39dcc9e7-c5a8-4fdd-9cf6-66300529dd5a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.749938] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efeb76c5-af18-41a6-aeed-2365bf1775c8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.775843] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Reconfiguring VM instance instance-0000001d to attach disk [datastore1] volume-2eb2f0fc-4130-44f7-bddb-eb01dbadccc4/volume-2eb2f0fc-4130-44f7-bddb-eb01dbadccc4.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 620.776358] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4befd106-ec09-42a6-a6bc-da488ccd7658 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.800025] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Waiting for the task: (returnval){ [ 620.800025] env[61728]: value = "task-463724" [ 620.800025] env[61728]: _type = "Task" [ 620.800025] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.806751] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-463724, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.809368] env[61728]: DEBUG nova.compute.manager [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 620.904910] env[61728]: DEBUG nova.network.neutron [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Successfully created port: 3196d005-b79e-4907-9b3c-59331e4d1e40 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 621.162904] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463723, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.252727] env[61728]: DEBUG nova.network.neutron [req-a051b3c2-bf3f-4be8-92dc-1773711c9de8 req-ef95b524-1cd6-4983-bc68-b17e6485a0c8 service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Updated VIF entry in instance network info cache for port 1915184e-ab7a-409c-a8d2-ccbdca5e942e. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 621.252727] env[61728]: DEBUG nova.network.neutron [req-a051b3c2-bf3f-4be8-92dc-1773711c9de8 req-ef95b524-1cd6-4983-bc68-b17e6485a0c8 service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Updating instance_info_cache with network_info: [{"id": "1915184e-ab7a-409c-a8d2-ccbdca5e942e", "address": "fa:16:3e:c4:37:f4", "network": {"id": "104934bf-5809-4ef5-adb6-970f5c45ea8e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1985848519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "667c26f99bea48cabb7d0bed8fa13b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7869cc8e-e58f-4fd6-88d7-85a18e43cd3a", "external-id": "nsx-vlan-transportzone-927", "segmentation_id": 927, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1915184e-ab", "ovs_interfaceid": "1915184e-ab7a-409c-a8d2-ccbdca5e942e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.328402] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-463724, 'name': ReconfigVM_Task, 'duration_secs': 0.368524} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.330492] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Reconfigured VM instance instance-0000001d to attach disk [datastore1] volume-2eb2f0fc-4130-44f7-bddb-eb01dbadccc4/volume-2eb2f0fc-4130-44f7-bddb-eb01dbadccc4.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 621.334998] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34a76f53-82c3-4d38-9f65-ff9cabcac5d9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.352289] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Waiting for the task: (returnval){ [ 621.352289] env[61728]: value = "task-463725" [ 621.352289] env[61728]: _type = "Task" [ 621.352289] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.368028] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-463725, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.490591] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4922bc4-2072-4143-880f-a0f9d91427f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.497987] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-949fe39a-5f48-4926-9aa7-5ba47c05f2d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.535012] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932669d1-717c-4363-9754-9002cea7c966 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.543345] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-280bcc6c-21fa-4400-9955-353bf40aa96a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.560107] env[61728]: DEBUG nova.compute.provider_tree [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.661615] env[61728]: DEBUG oslo_vmware.api [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463723, 'name': PowerOnVM_Task, 'duration_secs': 0.8717} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.661984] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 621.662260] env[61728]: INFO nova.compute.manager [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Took 8.34 seconds to spawn the instance on the hypervisor. [ 621.662653] env[61728]: DEBUG nova.compute.manager [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 621.663694] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36619afc-2a56-463a-861d-4dbeb9892b1a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.755099] env[61728]: DEBUG oslo_concurrency.lockutils [req-a051b3c2-bf3f-4be8-92dc-1773711c9de8 req-ef95b524-1cd6-4983-bc68-b17e6485a0c8 service nova] Releasing lock "refresh_cache-0f75629b-bdec-4593-aceb-cd0478972ccc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.831809] env[61728]: DEBUG nova.compute.manager [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 621.870274] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-463725, 'name': ReconfigVM_Task, 'duration_secs': 0.12668} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.870697] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-121946', 'volume_id': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'name': 'volume-2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0f75629b-bdec-4593-aceb-cd0478972ccc', 'attached_at': '', 'detached_at': '', 'volume_id': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'serial': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 621.871410] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7df16718-5e51-47f3-be28-bde30a9ddc56 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.878462] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Waiting for the task: (returnval){ [ 621.878462] env[61728]: value = "task-463726" [ 621.878462] env[61728]: _type = "Task" [ 621.878462] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.888703] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-463726, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.065953] env[61728]: DEBUG nova.scheduler.client.report [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.134131] env[61728]: DEBUG nova.virt.hardware [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 622.134131] env[61728]: DEBUG nova.virt.hardware [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 622.134131] env[61728]: DEBUG nova.virt.hardware [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 622.134690] env[61728]: DEBUG nova.virt.hardware [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 622.134690] env[61728]: DEBUG nova.virt.hardware [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 622.134690] env[61728]: DEBUG nova.virt.hardware [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 622.134690] env[61728]: DEBUG nova.virt.hardware [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 622.134690] env[61728]: DEBUG nova.virt.hardware [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 622.134962] env[61728]: DEBUG nova.virt.hardware [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 622.134962] env[61728]: DEBUG nova.virt.hardware [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 622.134962] env[61728]: DEBUG nova.virt.hardware [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 622.135239] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6740668b-5b49-47f4-9c89-c2731f0e6264 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.144266] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef37f57-8c33-4d37-b8d4-63f015a38bfb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.151166] env[61728]: DEBUG oslo_vmware.rw_handles [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52df4f08-9f18-c3e8-3b2b-3eb0b437a21e/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 622.151871] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62766d17-db2e-4372-8451-2541242df845 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.166212] env[61728]: DEBUG oslo_vmware.rw_handles [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52df4f08-9f18-c3e8-3b2b-3eb0b437a21e/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 622.166402] env[61728]: ERROR oslo_vmware.rw_handles [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52df4f08-9f18-c3e8-3b2b-3eb0b437a21e/disk-0.vmdk due to incomplete transfer. [ 622.166641] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-72988633-5b1f-4764-b0b7-c767487f3580 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.178710] env[61728]: DEBUG oslo_vmware.rw_handles [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52df4f08-9f18-c3e8-3b2b-3eb0b437a21e/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 622.178710] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Uploaded image 4b0aeaca-e578-41be-bb8c-6d3ee114d452 to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 622.181072] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 622.185437] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-83ffcb3e-f2bc-4575-81e7-b923dc56cff8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.185437] env[61728]: INFO nova.compute.manager [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Took 44.24 seconds to build instance. [ 622.192835] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 622.192835] env[61728]: value = "task-463727" [ 622.192835] env[61728]: _type = "Task" [ 622.192835] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.204729] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463727, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.390712] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-463726, 'name': Rename_Task, 'duration_secs': 0.136511} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.391270] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 622.391603] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b20666e5-4e28-48fa-81fb-65f331ba74dd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.398839] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Waiting for the task: (returnval){ [ 622.398839] env[61728]: value = "task-463728" [ 622.398839] env[61728]: _type = "Task" [ 622.398839] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.407944] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-463728, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.573125] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.773s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.574043] env[61728]: DEBUG nova.compute.manager [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 622.579053] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 33.538s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.579053] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.579053] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61728) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 622.580509] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.601s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.585064] env[61728]: INFO nova.compute.claims [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.592042] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7e30e4-1fc6-474b-a56c-f68146ddc973 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.599989] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87347d3-6960-46ca-be6e-a4ec581c512b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.619407] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2b7190-e30e-45d4-8966-32d1e224e6c9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.628640] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89619e18-1d81-4fea-b58d-42162ea37540 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.663517] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180273MB free_disk=114GB free_vcpus=48 pci_devices=None {{(pid=61728) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 622.664026] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.688072] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd3c06f3-bbd6-4439-8523-1dc6c5f66bfa tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.527s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.706522] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463727, 'name': Destroy_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.908722] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-463728, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.082402] env[61728]: DEBUG nova.network.neutron [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Successfully updated port: 3196d005-b79e-4907-9b3c-59331e4d1e40 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 623.090551] env[61728]: DEBUG nova.compute.utils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 623.091733] env[61728]: DEBUG nova.compute.manager [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 623.091958] env[61728]: DEBUG nova.network.neutron [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 623.180619] env[61728]: DEBUG nova.policy [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '38f64fd31712472a841f82c620d9340b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f5329f59d8bf43b2a855c4259387fccd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 623.191062] env[61728]: DEBUG nova.compute.manager [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 623.204138] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463727, 'name': Destroy_Task, 'duration_secs': 0.61414} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.204486] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Destroyed the VM [ 623.204765] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 623.205069] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a1de6fb1-ed74-4cd2-a5ae-8f918899408c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.213115] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 623.213115] env[61728]: value = "task-463729" [ 623.213115] env[61728]: _type = "Task" [ 623.213115] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.224028] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463729, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.377964] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.378589] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.413364] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-463728, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.578759] env[61728]: DEBUG nova.compute.manager [req-c34189c1-cbda-43c6-91b1-3cf81b27f75e req-3426acda-09b6-40a1-a0c8-a26ae0943ab0 service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Received event network-vif-plugged-3196d005-b79e-4907-9b3c-59331e4d1e40 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 623.578814] env[61728]: DEBUG oslo_concurrency.lockutils [req-c34189c1-cbda-43c6-91b1-3cf81b27f75e req-3426acda-09b6-40a1-a0c8-a26ae0943ab0 service nova] Acquiring lock "fbd521c8-9618-4b5c-839e-50d8631fa7b9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.579044] env[61728]: DEBUG oslo_concurrency.lockutils [req-c34189c1-cbda-43c6-91b1-3cf81b27f75e req-3426acda-09b6-40a1-a0c8-a26ae0943ab0 service nova] Lock "fbd521c8-9618-4b5c-839e-50d8631fa7b9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.579382] env[61728]: DEBUG oslo_concurrency.lockutils [req-c34189c1-cbda-43c6-91b1-3cf81b27f75e req-3426acda-09b6-40a1-a0c8-a26ae0943ab0 service nova] Lock "fbd521c8-9618-4b5c-839e-50d8631fa7b9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.579665] env[61728]: DEBUG nova.compute.manager [req-c34189c1-cbda-43c6-91b1-3cf81b27f75e req-3426acda-09b6-40a1-a0c8-a26ae0943ab0 service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] No waiting events found dispatching network-vif-plugged-3196d005-b79e-4907-9b3c-59331e4d1e40 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 623.579665] env[61728]: WARNING nova.compute.manager [req-c34189c1-cbda-43c6-91b1-3cf81b27f75e req-3426acda-09b6-40a1-a0c8-a26ae0943ab0 service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Received unexpected event network-vif-plugged-3196d005-b79e-4907-9b3c-59331e4d1e40 for instance with vm_state building and task_state spawning. [ 623.584439] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Acquiring lock "refresh_cache-fbd521c8-9618-4b5c-839e-50d8631fa7b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.584587] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Acquired lock "refresh_cache-fbd521c8-9618-4b5c-839e-50d8631fa7b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.584836] env[61728]: DEBUG nova.network.neutron [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 623.597376] env[61728]: DEBUG nova.compute.manager [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 623.670086] env[61728]: DEBUG nova.network.neutron [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Successfully created port: 9742f60c-5589-42f0-b3e4-a7d802b28b63 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.720444] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.728396] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463729, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.923730] env[61728]: DEBUG oslo_vmware.api [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-463728, 'name': PowerOnVM_Task, 'duration_secs': 1.111707} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.924189] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 623.924402] env[61728]: INFO nova.compute.manager [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Took 5.18 seconds to spawn the instance on the hypervisor. [ 623.924580] env[61728]: DEBUG nova.compute.manager [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 623.925940] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f4295a-af8d-4cbd-abd0-9a1fecbf6b44 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.120656] env[61728]: DEBUG nova.network.neutron [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.228038] env[61728]: DEBUG oslo_vmware.api [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463729, 'name': RemoveSnapshot_Task, 'duration_secs': 0.659766} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.228325] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 624.228610] env[61728]: INFO nova.compute.manager [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Took 14.63 seconds to snapshot the instance on the hypervisor. [ 624.239252] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4d9fe8-9f1e-46a6-b84c-51378cf731d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.249231] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc148f2-16d6-4686-808a-25efda0be0d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.289887] env[61728]: DEBUG nova.network.neutron [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Updating instance_info_cache with network_info: [{"id": "3196d005-b79e-4907-9b3c-59331e4d1e40", "address": "fa:16:3e:d1:5d:8a", "network": {"id": "527f7a39-789a-433b-9495-f13d2b52f0e0", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-189900763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99db3c843e8c4c3397cabba65ae510ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3196d005-b7", "ovs_interfaceid": "3196d005-b79e-4907-9b3c-59331e4d1e40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.291602] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac06c15-4738-47c6-ad29-e800bacabc0e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.300412] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccba1723-3f18-4de5-9665-655bbdc8f40c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.315821] env[61728]: DEBUG nova.compute.provider_tree [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.406605] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "fc3c1d93-b095-4d80-8d71-243b66d85a46" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.406865] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "fc3c1d93-b095-4d80-8d71-243b66d85a46" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.407131] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "fc3c1d93-b095-4d80-8d71-243b66d85a46-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.407392] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "fc3c1d93-b095-4d80-8d71-243b66d85a46-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.407457] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "fc3c1d93-b095-4d80-8d71-243b66d85a46-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.409940] env[61728]: INFO nova.compute.manager [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Terminating instance [ 624.411943] env[61728]: DEBUG nova.compute.manager [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 624.412293] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 624.413302] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703cd39a-3f0e-4aaf-86db-35e5cea3e7d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.422811] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 624.422927] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05a54102-5eec-4aad-a646-9f2a8b75f866 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.455883] env[61728]: INFO nova.compute.manager [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Took 41.82 seconds to build instance. [ 624.488874] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 624.489105] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 624.489565] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleting the datastore file [datastore1] fc3c1d93-b095-4d80-8d71-243b66d85a46 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 624.489936] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec1cf68d-7b54-4f73-8b67-e69538eadaad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.497342] env[61728]: DEBUG oslo_vmware.api [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 624.497342] env[61728]: value = "task-463731" [ 624.497342] env[61728]: _type = "Task" [ 624.497342] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.505976] env[61728]: DEBUG oslo_vmware.api [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463731, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.611693] env[61728]: DEBUG nova.compute.manager [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 624.633806] env[61728]: DEBUG nova.virt.hardware [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 624.634211] env[61728]: DEBUG nova.virt.hardware [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 624.634387] env[61728]: DEBUG nova.virt.hardware [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 624.634641] env[61728]: DEBUG nova.virt.hardware [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 624.634840] env[61728]: DEBUG nova.virt.hardware [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 624.635034] env[61728]: DEBUG nova.virt.hardware [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 624.635311] env[61728]: DEBUG nova.virt.hardware [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 624.635554] env[61728]: DEBUG nova.virt.hardware [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 624.635800] env[61728]: DEBUG nova.virt.hardware [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 624.636010] env[61728]: DEBUG nova.virt.hardware [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 624.636238] env[61728]: DEBUG nova.virt.hardware [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 624.637160] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1269043-c8e8-4ab6-bdcd-8e8493787dab {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.645545] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6477b58a-5471-494c-8bb9-ffd5bea52700 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.736589] env[61728]: DEBUG nova.compute.manager [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Instance disappeared during snapshot {{(pid=61728) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 624.750192] env[61728]: DEBUG nova.compute.manager [None req-33fdf956-c312-43be-a83e-5e3460c650d8 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Image not found during clean up 4b0aeaca-e578-41be-bb8c-6d3ee114d452 {{(pid=61728) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 624.795931] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Releasing lock "refresh_cache-fbd521c8-9618-4b5c-839e-50d8631fa7b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.796371] env[61728]: DEBUG nova.compute.manager [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Instance network_info: |[{"id": "3196d005-b79e-4907-9b3c-59331e4d1e40", "address": "fa:16:3e:d1:5d:8a", "network": {"id": "527f7a39-789a-433b-9495-f13d2b52f0e0", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-189900763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99db3c843e8c4c3397cabba65ae510ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3196d005-b7", "ovs_interfaceid": "3196d005-b79e-4907-9b3c-59331e4d1e40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 624.796887] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:5d:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fb6e5b3-7be7-4018-8d9e-2133d926178c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3196d005-b79e-4907-9b3c-59331e4d1e40', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 624.805308] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Creating folder: Project (99db3c843e8c4c3397cabba65ae510ca). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 624.806130] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dcfb5c1e-03a3-4e8a-b1ec-1c624e0e4b55 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.817208] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Created folder: Project (99db3c843e8c4c3397cabba65ae510ca) in parent group-v121913. [ 624.817538] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Creating folder: Instances. Parent ref: group-v122008. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 624.818541] env[61728]: DEBUG nova.scheduler.client.report [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.822200] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0bded17c-1e83-4bcf-b186-33169cb1e2d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.832626] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Created folder: Instances in parent group-v122008. [ 624.832626] env[61728]: DEBUG oslo.service.loopingcall [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 624.832849] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 624.833339] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e9b4abc9-4309-429d-8bfc-19bde029b293 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.855149] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 624.855149] env[61728]: value = "task-463734" [ 624.855149] env[61728]: _type = "Task" [ 624.855149] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.865830] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463734, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.957688] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ac60923-508d-4f57-9a7c-f90f64e2b992 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Lock "0f75629b-bdec-4593-aceb-cd0478972ccc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.469s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.008460] env[61728]: DEBUG oslo_vmware.api [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463731, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.244122} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.009130] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 625.009560] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 625.009933] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 625.010280] env[61728]: INFO nova.compute.manager [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Took 0.60 seconds to destroy the instance on the hypervisor. [ 625.010650] env[61728]: DEBUG oslo.service.loopingcall [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.010972] env[61728]: DEBUG nova.compute.manager [-] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.011172] env[61728]: DEBUG nova.network.neutron [-] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 625.326736] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.748s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.327345] env[61728]: DEBUG nova.compute.manager [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 625.338020] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.489s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.338020] env[61728]: DEBUG nova.objects.instance [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Lazy-loading 'resources' on Instance uuid aec920d1-12fb-4d41-9663-b4c2b7e142d7 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 625.369538] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463734, 'name': CreateVM_Task, 'duration_secs': 0.460326} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.370579] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 625.372084] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.372084] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.372084] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 625.372084] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8b73866-9043-4e96-afac-cc340bd17976 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.378921] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Waiting for the task: (returnval){ [ 625.378921] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524ae2ba-0e7b-a9f1-8f52-0d276f916cab" [ 625.378921] env[61728]: _type = "Task" [ 625.378921] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.390158] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524ae2ba-0e7b-a9f1-8f52-0d276f916cab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.418747] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "c14034b5-330b-4fb1-af31-d530e3231992" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.419014] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "c14034b5-330b-4fb1-af31-d530e3231992" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.461356] env[61728]: DEBUG nova.compute.manager [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.699413] env[61728]: DEBUG nova.compute.manager [req-3b1f8fab-7bde-464a-98a5-fb81ed02b461 req-4eadcfe0-0ee5-41e2-98b6-9e0aff5ce919 service nova] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Received event network-vif-deleted-7a9a7fa8-9c61-46be-a49a-8dd67bce2d22 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 625.699413] env[61728]: INFO nova.compute.manager [req-3b1f8fab-7bde-464a-98a5-fb81ed02b461 req-4eadcfe0-0ee5-41e2-98b6-9e0aff5ce919 service nova] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Neutron deleted interface 7a9a7fa8-9c61-46be-a49a-8dd67bce2d22; detaching it from the instance and deleting it from the info cache [ 625.699413] env[61728]: DEBUG nova.network.neutron [req-3b1f8fab-7bde-464a-98a5-fb81ed02b461 req-4eadcfe0-0ee5-41e2-98b6-9e0aff5ce919 service nova] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.720047] env[61728]: DEBUG nova.compute.manager [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Received event network-changed-3196d005-b79e-4907-9b3c-59331e4d1e40 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 625.720371] env[61728]: DEBUG nova.compute.manager [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Refreshing instance network info cache due to event network-changed-3196d005-b79e-4907-9b3c-59331e4d1e40. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 625.720673] env[61728]: DEBUG oslo_concurrency.lockutils [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] Acquiring lock "refresh_cache-fbd521c8-9618-4b5c-839e-50d8631fa7b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.721312] env[61728]: DEBUG oslo_concurrency.lockutils [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] Acquired lock "refresh_cache-fbd521c8-9618-4b5c-839e-50d8631fa7b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.721312] env[61728]: DEBUG nova.network.neutron [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Refreshing network info cache for port 3196d005-b79e-4907-9b3c-59331e4d1e40 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 625.837644] env[61728]: DEBUG nova.compute.utils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 625.839353] env[61728]: DEBUG nova.compute.manager [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 625.843024] env[61728]: DEBUG nova.network.neutron [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 625.895375] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524ae2ba-0e7b-a9f1-8f52-0d276f916cab, 'name': SearchDatastore_Task, 'duration_secs': 0.013518} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.895779] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.896070] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 625.896356] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.896549] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.897300] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 625.898397] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7b165fcd-80d8-43c4-bbfd-231f7f1f41f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.913923] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 625.914146] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 625.914925] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f908299b-d342-4ded-b696-f170ad7f9622 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.921116] env[61728]: DEBUG nova.policy [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fa90ed28524347fdacc6126b82d69596', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'db60141027bf4fc7bcb9ae44ab4e77ed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 625.927187] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Waiting for the task: (returnval){ [ 625.927187] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529dcadf-1b97-f584-6e17-9cb0abb0d917" [ 625.927187] env[61728]: _type = "Task" [ 625.927187] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.938458] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529dcadf-1b97-f584-6e17-9cb0abb0d917, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.975059] env[61728]: DEBUG nova.network.neutron [-] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.999293] env[61728]: DEBUG nova.network.neutron [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Successfully updated port: 9742f60c-5589-42f0-b3e4-a7d802b28b63 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 626.001388] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.204994] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-07c210f9-b0d4-4503-b76f-b502954bc14d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.215670] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e69bb4ef-0d30-4d39-9b75-4825368980cf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.254725] env[61728]: DEBUG nova.compute.manager [req-3b1f8fab-7bde-464a-98a5-fb81ed02b461 req-4eadcfe0-0ee5-41e2-98b6-9e0aff5ce919 service nova] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Detach interface failed, port_id=7a9a7fa8-9c61-46be-a49a-8dd67bce2d22, reason: Instance fc3c1d93-b095-4d80-8d71-243b66d85a46 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 626.272466] env[61728]: DEBUG nova.network.neutron [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Successfully created port: a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.344814] env[61728]: DEBUG nova.compute.manager [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 626.444379] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529dcadf-1b97-f584-6e17-9cb0abb0d917, 'name': SearchDatastore_Task, 'duration_secs': 0.031521} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.450208] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-caa56435-065c-45ab-9883-25702f9738f6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.455080] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Waiting for the task: (returnval){ [ 626.455080] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52147413-2354-b1a3-d150-aa992537cc1d" [ 626.455080] env[61728]: _type = "Task" [ 626.455080] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.466205] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52147413-2354-b1a3-d150-aa992537cc1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.479887] env[61728]: INFO nova.compute.manager [-] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Took 1.47 seconds to deallocate network for instance. [ 626.507829] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "refresh_cache-1ef96606-d960-4283-b372-e2bad47f9367" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.508171] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "refresh_cache-1ef96606-d960-4283-b372-e2bad47f9367" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.512019] env[61728]: DEBUG nova.network.neutron [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 626.523106] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da863ec-5ff5-43f0-b855-d698f03a3c72 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.532608] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f760de62-6c5f-408c-8426-f88cc67a70b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.573193] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe0d4fe-b133-4482-ba83-4f411c1931e8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.585275] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-107ca1e4-2431-47d9-ba7a-8df56f5c2f40 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.600664] env[61728]: DEBUG nova.compute.provider_tree [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.647701] env[61728]: DEBUG nova.network.neutron [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Updated VIF entry in instance network info cache for port 3196d005-b79e-4907-9b3c-59331e4d1e40. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 626.647851] env[61728]: DEBUG nova.network.neutron [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Updating instance_info_cache with network_info: [{"id": "3196d005-b79e-4907-9b3c-59331e4d1e40", "address": "fa:16:3e:d1:5d:8a", "network": {"id": "527f7a39-789a-433b-9495-f13d2b52f0e0", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-189900763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99db3c843e8c4c3397cabba65ae510ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3196d005-b7", "ovs_interfaceid": "3196d005-b79e-4907-9b3c-59331e4d1e40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.968388] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52147413-2354-b1a3-d150-aa992537cc1d, 'name': SearchDatastore_Task, 'duration_secs': 0.013336} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.968388] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.968388] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] fbd521c8-9618-4b5c-839e-50d8631fa7b9/fbd521c8-9618-4b5c-839e-50d8631fa7b9.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 626.968388] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d417dbb-275d-4cd6-90d5-50c475a98b23 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.976735] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Waiting for the task: (returnval){ [ 626.976735] env[61728]: value = "task-463735" [ 626.976735] env[61728]: _type = "Task" [ 626.976735] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.983040] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463735, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.986933] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.057309] env[61728]: DEBUG nova.network.neutron [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.103808] env[61728]: DEBUG nova.scheduler.client.report [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.156846] env[61728]: DEBUG oslo_concurrency.lockutils [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] Releasing lock "refresh_cache-fbd521c8-9618-4b5c-839e-50d8631fa7b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.157159] env[61728]: DEBUG nova.compute.manager [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Received event network-changed-ccffc76c-1df1-4ba7-824a-5c175febeca3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 627.157336] env[61728]: DEBUG nova.compute.manager [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Refreshing instance network info cache due to event network-changed-ccffc76c-1df1-4ba7-824a-5c175febeca3. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 627.157556] env[61728]: DEBUG oslo_concurrency.lockutils [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] Acquiring lock "refresh_cache-c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.157702] env[61728]: DEBUG oslo_concurrency.lockutils [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] Acquired lock "refresh_cache-c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.157868] env[61728]: DEBUG nova.network.neutron [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Refreshing network info cache for port ccffc76c-1df1-4ba7-824a-5c175febeca3 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 627.322300] env[61728]: DEBUG nova.network.neutron [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Updating instance_info_cache with network_info: [{"id": "9742f60c-5589-42f0-b3e4-a7d802b28b63", "address": "fa:16:3e:81:1c:ba", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9742f60c-55", "ovs_interfaceid": "9742f60c-5589-42f0-b3e4-a7d802b28b63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.353150] env[61728]: DEBUG nova.compute.manager [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 627.381393] env[61728]: DEBUG nova.virt.hardware [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 627.381714] env[61728]: DEBUG nova.virt.hardware [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 627.381895] env[61728]: DEBUG nova.virt.hardware [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 627.382086] env[61728]: DEBUG nova.virt.hardware [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 627.382238] env[61728]: DEBUG nova.virt.hardware [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 627.382431] env[61728]: DEBUG nova.virt.hardware [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 627.382668] env[61728]: DEBUG nova.virt.hardware [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 627.382877] env[61728]: DEBUG nova.virt.hardware [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 627.383104] env[61728]: DEBUG nova.virt.hardware [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 627.383290] env[61728]: DEBUG nova.virt.hardware [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 627.383484] env[61728]: DEBUG nova.virt.hardware [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 627.384412] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3949bb93-f6b7-4acd-8488-45670001c0e9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.392562] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcafa7df-7215-435f-9b67-2beeda8b0b75 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.485139] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463735, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.609060] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.274s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.611527] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.250s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.613864] env[61728]: DEBUG nova.objects.instance [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Lazy-loading 'resources' on Instance uuid 2c76f45a-b87c-4c3b-9bf9-ba5d67265437 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 627.632656] env[61728]: INFO nova.scheduler.client.report [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Deleted allocations for instance aec920d1-12fb-4d41-9663-b4c2b7e142d7 [ 627.825740] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "refresh_cache-1ef96606-d960-4283-b372-e2bad47f9367" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.826331] env[61728]: DEBUG nova.compute.manager [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Instance network_info: |[{"id": "9742f60c-5589-42f0-b3e4-a7d802b28b63", "address": "fa:16:3e:81:1c:ba", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9742f60c-55", "ovs_interfaceid": "9742f60c-5589-42f0-b3e4-a7d802b28b63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 627.826822] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:1c:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c492f5cc-7ae0-4cab-823c-0d5dd8c60b26', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9742f60c-5589-42f0-b3e4-a7d802b28b63', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 627.835896] env[61728]: DEBUG oslo.service.loopingcall [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 627.836566] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 627.836872] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a2d2251-0866-40a0-b536-6aa5c59c29f6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.857499] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 627.857499] env[61728]: value = "task-463736" [ 627.857499] env[61728]: _type = "Task" [ 627.857499] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.862938] env[61728]: DEBUG nova.network.neutron [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Successfully updated port: a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 627.872619] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463736, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.965371] env[61728]: DEBUG nova.compute.manager [req-c62aedd6-0b2c-4d53-9601-757aa3e61d17 req-9a19cc6e-542d-4801-bc4e-14f11279998e service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Received event network-changed-1915184e-ab7a-409c-a8d2-ccbdca5e942e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 627.965568] env[61728]: DEBUG nova.compute.manager [req-c62aedd6-0b2c-4d53-9601-757aa3e61d17 req-9a19cc6e-542d-4801-bc4e-14f11279998e service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Refreshing instance network info cache due to event network-changed-1915184e-ab7a-409c-a8d2-ccbdca5e942e. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 627.965788] env[61728]: DEBUG oslo_concurrency.lockutils [req-c62aedd6-0b2c-4d53-9601-757aa3e61d17 req-9a19cc6e-542d-4801-bc4e-14f11279998e service nova] Acquiring lock "refresh_cache-0f75629b-bdec-4593-aceb-cd0478972ccc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.965935] env[61728]: DEBUG oslo_concurrency.lockutils [req-c62aedd6-0b2c-4d53-9601-757aa3e61d17 req-9a19cc6e-542d-4801-bc4e-14f11279998e service nova] Acquired lock "refresh_cache-0f75629b-bdec-4593-aceb-cd0478972ccc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.966429] env[61728]: DEBUG nova.network.neutron [req-c62aedd6-0b2c-4d53-9601-757aa3e61d17 req-9a19cc6e-542d-4801-bc4e-14f11279998e service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Refreshing network info cache for port 1915184e-ab7a-409c-a8d2-ccbdca5e942e {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 627.986690] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463735, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557556} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.987134] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] fbd521c8-9618-4b5c-839e-50d8631fa7b9/fbd521c8-9618-4b5c-839e-50d8631fa7b9.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 627.987134] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 627.987366] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bcd74543-8ac5-4a34-97b8-66515b5b17e0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.994545] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Waiting for the task: (returnval){ [ 627.994545] env[61728]: value = "task-463737" [ 627.994545] env[61728]: _type = "Task" [ 627.994545] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.003228] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463737, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.046661] env[61728]: DEBUG nova.network.neutron [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Updated VIF entry in instance network info cache for port ccffc76c-1df1-4ba7-824a-5c175febeca3. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 628.047062] env[61728]: DEBUG nova.network.neutron [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Updating instance_info_cache with network_info: [{"id": "ccffc76c-1df1-4ba7-824a-5c175febeca3", "address": "fa:16:3e:33:14:5f", "network": {"id": "5d1e643d-f8a4-4302-9920-bb661df4632d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1891316321-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fc16f1d1b144286b180cd7c8f7588fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapccffc76c-1d", "ovs_interfaceid": "ccffc76c-1df1-4ba7-824a-5c175febeca3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.147931] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07a16969-abb0-4855-98f7-156a0052c97b tempest-TenantUsagesTestJSON-1755717129 tempest-TenantUsagesTestJSON-1755717129-project-member] Lock "aec920d1-12fb-4d41-9663-b4c2b7e142d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.241s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.179278] env[61728]: DEBUG nova.compute.manager [req-00383fee-805a-411f-9bef-a1de9bbc34f6 req-367fe8cd-0606-49eb-af18-130c7516cefe service nova] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Received event network-changed-9742f60c-5589-42f0-b3e4-a7d802b28b63 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 628.179527] env[61728]: DEBUG nova.compute.manager [req-00383fee-805a-411f-9bef-a1de9bbc34f6 req-367fe8cd-0606-49eb-af18-130c7516cefe service nova] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Refreshing instance network info cache due to event network-changed-9742f60c-5589-42f0-b3e4-a7d802b28b63. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 628.179816] env[61728]: DEBUG oslo_concurrency.lockutils [req-00383fee-805a-411f-9bef-a1de9bbc34f6 req-367fe8cd-0606-49eb-af18-130c7516cefe service nova] Acquiring lock "refresh_cache-1ef96606-d960-4283-b372-e2bad47f9367" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.179975] env[61728]: DEBUG oslo_concurrency.lockutils [req-00383fee-805a-411f-9bef-a1de9bbc34f6 req-367fe8cd-0606-49eb-af18-130c7516cefe service nova] Acquired lock "refresh_cache-1ef96606-d960-4283-b372-e2bad47f9367" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.180161] env[61728]: DEBUG nova.network.neutron [req-00383fee-805a-411f-9bef-a1de9bbc34f6 req-367fe8cd-0606-49eb-af18-130c7516cefe service nova] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Refreshing network info cache for port 9742f60c-5589-42f0-b3e4-a7d802b28b63 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 628.379285] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquiring lock "refresh_cache-7a60fcd0-01b0-40bc-93e7-6f5ce03232d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.379285] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquired lock "refresh_cache-7a60fcd0-01b0-40bc-93e7-6f5ce03232d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.379285] env[61728]: DEBUG nova.network.neutron [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 628.380716] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463736, 'name': CreateVM_Task, 'duration_secs': 0.48464} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.386597] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 628.387301] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.387471] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.387829] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 628.388053] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9e11984-6fba-460f-80b4-4effb13cb9a3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.394934] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 628.394934] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525a4139-8459-6bd6-e1a6-0935fc94a9c6" [ 628.394934] env[61728]: _type = "Task" [ 628.394934] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.410217] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525a4139-8459-6bd6-e1a6-0935fc94a9c6, 'name': SearchDatastore_Task, 'duration_secs': 0.011863} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.412054] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.412054] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 628.412054] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.412054] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.412411] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 628.412411] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db01fad5-432c-4dde-b2fe-21de544461ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.429972] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 628.429972] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 628.431058] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc6599f8-bcd1-4b89-adf0-242407df1c33 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.439774] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 628.439774] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521b64f9-9686-2a20-ae52-49f73fcced21" [ 628.439774] env[61728]: _type = "Task" [ 628.439774] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.447986] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521b64f9-9686-2a20-ae52-49f73fcced21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.510758] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463737, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090187} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.510758] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 628.510758] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed7cd31-7de6-4ed5-a4e6-d942f32381a0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.533483] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Reconfiguring VM instance instance-0000001e to attach disk [datastore1] fbd521c8-9618-4b5c-839e-50d8631fa7b9/fbd521c8-9618-4b5c-839e-50d8631fa7b9.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 628.538739] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c383b11-b21f-4742-9317-ecacc7acbc1d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.553549] env[61728]: DEBUG oslo_concurrency.lockutils [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] Releasing lock "refresh_cache-c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.553822] env[61728]: DEBUG nova.compute.manager [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Received event network-changed-ccffc76c-1df1-4ba7-824a-5c175febeca3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 628.553995] env[61728]: DEBUG nova.compute.manager [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Refreshing instance network info cache due to event network-changed-ccffc76c-1df1-4ba7-824a-5c175febeca3. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 628.554221] env[61728]: DEBUG oslo_concurrency.lockutils [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] Acquiring lock "refresh_cache-c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.554368] env[61728]: DEBUG oslo_concurrency.lockutils [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] Acquired lock "refresh_cache-c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.554540] env[61728]: DEBUG nova.network.neutron [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Refreshing network info cache for port ccffc76c-1df1-4ba7-824a-5c175febeca3 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 628.561858] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Waiting for the task: (returnval){ [ 628.561858] env[61728]: value = "task-463738" [ 628.561858] env[61728]: _type = "Task" [ 628.561858] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.573817] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463738, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.681106] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceea1c82-cd11-438b-8ae2-9a9b482e0214 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.694021] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd3b2d6-6d2b-4c91-acd0-450288f16a84 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.728415] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-364da491-7896-4e42-847a-05761103cdee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.736898] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540c4130-24f8-4e9a-8384-11e7321ce300 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.753455] env[61728]: DEBUG nova.compute.provider_tree [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.876999] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c53eb3-6ff0-8632-fd43-28918162934d/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 628.878009] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e4feb3-ce78-452f-bea9-2b5f859f9c74 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.885692] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c53eb3-6ff0-8632-fd43-28918162934d/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 628.885860] env[61728]: ERROR oslo_vmware.rw_handles [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c53eb3-6ff0-8632-fd43-28918162934d/disk-0.vmdk due to incomplete transfer. [ 628.888459] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-fd4f2d82-c252-414e-9657-255b371a6034 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.899964] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c53eb3-6ff0-8632-fd43-28918162934d/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 628.900134] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Uploaded image a203be0f-1f36-4177-9319-6056c97931b5 to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 628.904060] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 628.904060] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-47707b24-d65e-46f1-bb59-ebb9da005ba7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.907546] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 628.907546] env[61728]: value = "task-463739" [ 628.907546] env[61728]: _type = "Task" [ 628.907546] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.916360] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463739, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.945562] env[61728]: DEBUG nova.network.neutron [req-00383fee-805a-411f-9bef-a1de9bbc34f6 req-367fe8cd-0606-49eb-af18-130c7516cefe service nova] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Updated VIF entry in instance network info cache for port 9742f60c-5589-42f0-b3e4-a7d802b28b63. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 628.945997] env[61728]: DEBUG nova.network.neutron [req-00383fee-805a-411f-9bef-a1de9bbc34f6 req-367fe8cd-0606-49eb-af18-130c7516cefe service nova] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Updating instance_info_cache with network_info: [{"id": "9742f60c-5589-42f0-b3e4-a7d802b28b63", "address": "fa:16:3e:81:1c:ba", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9742f60c-55", "ovs_interfaceid": "9742f60c-5589-42f0-b3e4-a7d802b28b63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.954665] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521b64f9-9686-2a20-ae52-49f73fcced21, 'name': SearchDatastore_Task, 'duration_secs': 0.021166} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.955536] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6e12b44-fd97-45a8-b371-5ac042216f68 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.960870] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 628.960870] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e1d709-7506-439b-ff53-b95d5fff889a" [ 628.960870] env[61728]: _type = "Task" [ 628.960870] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.969920] env[61728]: DEBUG nova.network.neutron [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.975923] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e1d709-7506-439b-ff53-b95d5fff889a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.012094] env[61728]: DEBUG nova.network.neutron [req-c62aedd6-0b2c-4d53-9601-757aa3e61d17 req-9a19cc6e-542d-4801-bc4e-14f11279998e service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Updated VIF entry in instance network info cache for port 1915184e-ab7a-409c-a8d2-ccbdca5e942e. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 629.012506] env[61728]: DEBUG nova.network.neutron [req-c62aedd6-0b2c-4d53-9601-757aa3e61d17 req-9a19cc6e-542d-4801-bc4e-14f11279998e service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Updating instance_info_cache with network_info: [{"id": "1915184e-ab7a-409c-a8d2-ccbdca5e942e", "address": "fa:16:3e:c4:37:f4", "network": {"id": "104934bf-5809-4ef5-adb6-970f5c45ea8e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1985848519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "667c26f99bea48cabb7d0bed8fa13b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7869cc8e-e58f-4fd6-88d7-85a18e43cd3a", "external-id": "nsx-vlan-transportzone-927", "segmentation_id": 927, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1915184e-ab", "ovs_interfaceid": "1915184e-ab7a-409c-a8d2-ccbdca5e942e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.072193] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.125636] env[61728]: DEBUG nova.network.neutron [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Updating instance_info_cache with network_info: [{"id": "a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7", "address": "fa:16:3e:6d:36:9d", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.216", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2de5ad6-df", "ovs_interfaceid": "a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.257437] env[61728]: DEBUG nova.scheduler.client.report [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.418853] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463739, 'name': Destroy_Task, 'duration_secs': 0.400008} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.419183] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Destroyed the VM [ 629.419487] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 629.419715] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1fc9bde5-6baf-4d31-9f73-81e637457d51 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.426390] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 629.426390] env[61728]: value = "task-463740" [ 629.426390] env[61728]: _type = "Task" [ 629.426390] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.435894] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463740, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.448734] env[61728]: DEBUG oslo_concurrency.lockutils [req-00383fee-805a-411f-9bef-a1de9bbc34f6 req-367fe8cd-0606-49eb-af18-130c7516cefe service nova] Releasing lock "refresh_cache-1ef96606-d960-4283-b372-e2bad47f9367" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.458293] env[61728]: DEBUG nova.network.neutron [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Updated VIF entry in instance network info cache for port ccffc76c-1df1-4ba7-824a-5c175febeca3. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 629.458662] env[61728]: DEBUG nova.network.neutron [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Updating instance_info_cache with network_info: [{"id": "ccffc76c-1df1-4ba7-824a-5c175febeca3", "address": "fa:16:3e:33:14:5f", "network": {"id": "5d1e643d-f8a4-4302-9920-bb661df4632d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1891316321-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fc16f1d1b144286b180cd7c8f7588fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapccffc76c-1d", "ovs_interfaceid": "ccffc76c-1df1-4ba7-824a-5c175febeca3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.471204] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e1d709-7506-439b-ff53-b95d5fff889a, 'name': SearchDatastore_Task, 'duration_secs': 0.012296} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.472474] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.472778] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 1ef96606-d960-4283-b372-e2bad47f9367/1ef96606-d960-4283-b372-e2bad47f9367.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 629.473079] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ee4aa1ad-1f4f-4d34-87b6-58afdd1719cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.481190] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 629.481190] env[61728]: value = "task-463741" [ 629.481190] env[61728]: _type = "Task" [ 629.481190] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.492479] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463741, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.515175] env[61728]: DEBUG oslo_concurrency.lockutils [req-c62aedd6-0b2c-4d53-9601-757aa3e61d17 req-9a19cc6e-542d-4801-bc4e-14f11279998e service nova] Releasing lock "refresh_cache-0f75629b-bdec-4593-aceb-cd0478972ccc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.572391] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463738, 'name': ReconfigVM_Task, 'duration_secs': 0.711095} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.572776] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Reconfigured VM instance instance-0000001e to attach disk [datastore1] fbd521c8-9618-4b5c-839e-50d8631fa7b9/fbd521c8-9618-4b5c-839e-50d8631fa7b9.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 629.573653] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6cec46b1-78cf-441e-8cc7-af2e7abba580 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.579566] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Waiting for the task: (returnval){ [ 629.579566] env[61728]: value = "task-463742" [ 629.579566] env[61728]: _type = "Task" [ 629.579566] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.588385] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463742, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.631024] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Releasing lock "refresh_cache-7a60fcd0-01b0-40bc-93e7-6f5ce03232d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.631024] env[61728]: DEBUG nova.compute.manager [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Instance network_info: |[{"id": "a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7", "address": "fa:16:3e:6d:36:9d", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.216", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2de5ad6-df", "ovs_interfaceid": "a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 629.631330] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:36:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '457c42cd-4ddb-4374-923e-d419b7f6eaff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 629.638688] env[61728]: DEBUG oslo.service.loopingcall [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 629.639242] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 629.639830] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0ee94b3a-7dde-42eb-9584-028ec2e12615 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.661755] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 629.661755] env[61728]: value = "task-463743" [ 629.661755] env[61728]: _type = "Task" [ 629.661755] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.672554] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463743, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.767185] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.155s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.770254] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.323s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.772363] env[61728]: INFO nova.compute.claims [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 629.798512] env[61728]: INFO nova.scheduler.client.report [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Deleted allocations for instance 2c76f45a-b87c-4c3b-9bf9-ba5d67265437 [ 629.939731] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463740, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.961650] env[61728]: DEBUG oslo_concurrency.lockutils [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] Releasing lock "refresh_cache-c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.961963] env[61728]: DEBUG nova.compute.manager [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Received event network-vif-plugged-9742f60c-5589-42f0-b3e4-a7d802b28b63 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 629.962187] env[61728]: DEBUG oslo_concurrency.lockutils [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] Acquiring lock "1ef96606-d960-4283-b372-e2bad47f9367-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.962409] env[61728]: DEBUG oslo_concurrency.lockutils [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] Lock "1ef96606-d960-4283-b372-e2bad47f9367-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.962575] env[61728]: DEBUG oslo_concurrency.lockutils [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] Lock "1ef96606-d960-4283-b372-e2bad47f9367-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.962750] env[61728]: DEBUG nova.compute.manager [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] No waiting events found dispatching network-vif-plugged-9742f60c-5589-42f0-b3e4-a7d802b28b63 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 629.962982] env[61728]: WARNING nova.compute.manager [req-49b5e9d0-54db-46b3-870c-baeb6cf7034f req-3a5d7508-1e11-4800-af0b-335b9b06a0b6 service nova] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Received unexpected event network-vif-plugged-9742f60c-5589-42f0-b3e4-a7d802b28b63 for instance with vm_state building and task_state spawning. [ 629.990445] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463741, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493452} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.991197] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 1ef96606-d960-4283-b372-e2bad47f9367/1ef96606-d960-4283-b372-e2bad47f9367.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 629.991197] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 629.991197] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a72b9258-d56b-4109-82ce-2210857acd14 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.998358] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 629.998358] env[61728]: value = "task-463744" [ 629.998358] env[61728]: _type = "Task" [ 629.998358] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.008053] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463744, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.091517] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463742, 'name': Rename_Task, 'duration_secs': 0.207479} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.091991] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 630.092281] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c658346e-2d78-4620-8cd2-f0c8ca5b36a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.099058] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Waiting for the task: (returnval){ [ 630.099058] env[61728]: value = "task-463745" [ 630.099058] env[61728]: _type = "Task" [ 630.099058] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.107126] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463745, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.174259] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463743, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.204524] env[61728]: DEBUG nova.compute.manager [req-5cbb4239-5118-4085-adee-61f2f913cbef req-0f33487f-2039-49e4-88a9-b4a48c53e100 service nova] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Received event network-vif-plugged-a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 630.204785] env[61728]: DEBUG oslo_concurrency.lockutils [req-5cbb4239-5118-4085-adee-61f2f913cbef req-0f33487f-2039-49e4-88a9-b4a48c53e100 service nova] Acquiring lock "7a60fcd0-01b0-40bc-93e7-6f5ce03232d4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.205052] env[61728]: DEBUG oslo_concurrency.lockutils [req-5cbb4239-5118-4085-adee-61f2f913cbef req-0f33487f-2039-49e4-88a9-b4a48c53e100 service nova] Lock "7a60fcd0-01b0-40bc-93e7-6f5ce03232d4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.205253] env[61728]: DEBUG oslo_concurrency.lockutils [req-5cbb4239-5118-4085-adee-61f2f913cbef req-0f33487f-2039-49e4-88a9-b4a48c53e100 service nova] Lock "7a60fcd0-01b0-40bc-93e7-6f5ce03232d4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.205488] env[61728]: DEBUG nova.compute.manager [req-5cbb4239-5118-4085-adee-61f2f913cbef req-0f33487f-2039-49e4-88a9-b4a48c53e100 service nova] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] No waiting events found dispatching network-vif-plugged-a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 630.205722] env[61728]: WARNING nova.compute.manager [req-5cbb4239-5118-4085-adee-61f2f913cbef req-0f33487f-2039-49e4-88a9-b4a48c53e100 service nova] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Received unexpected event network-vif-plugged-a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7 for instance with vm_state building and task_state spawning. [ 630.206059] env[61728]: DEBUG nova.compute.manager [req-5cbb4239-5118-4085-adee-61f2f913cbef req-0f33487f-2039-49e4-88a9-b4a48c53e100 service nova] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Received event network-changed-a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 630.206462] env[61728]: DEBUG nova.compute.manager [req-5cbb4239-5118-4085-adee-61f2f913cbef req-0f33487f-2039-49e4-88a9-b4a48c53e100 service nova] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Refreshing instance network info cache due to event network-changed-a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 630.206676] env[61728]: DEBUG oslo_concurrency.lockutils [req-5cbb4239-5118-4085-adee-61f2f913cbef req-0f33487f-2039-49e4-88a9-b4a48c53e100 service nova] Acquiring lock "refresh_cache-7a60fcd0-01b0-40bc-93e7-6f5ce03232d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.206784] env[61728]: DEBUG oslo_concurrency.lockutils [req-5cbb4239-5118-4085-adee-61f2f913cbef req-0f33487f-2039-49e4-88a9-b4a48c53e100 service nova] Acquired lock "refresh_cache-7a60fcd0-01b0-40bc-93e7-6f5ce03232d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.207362] env[61728]: DEBUG nova.network.neutron [req-5cbb4239-5118-4085-adee-61f2f913cbef req-0f33487f-2039-49e4-88a9-b4a48c53e100 service nova] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Refreshing network info cache for port a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 630.308274] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00da756d-a270-45a0-890d-7053be2c62dd tempest-ServerDiagnosticsV248Test-1436453890 tempest-ServerDiagnosticsV248Test-1436453890-project-member] Lock "2c76f45a-b87c-4c3b-9bf9-ba5d67265437" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.840s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.440145] env[61728]: DEBUG oslo_vmware.api [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463740, 'name': RemoveSnapshot_Task, 'duration_secs': 0.950282} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.440370] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 630.440609] env[61728]: INFO nova.compute.manager [None req-b86b8cce-9f2f-4903-b9cf-e78029551185 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Took 15.06 seconds to snapshot the instance on the hypervisor. [ 630.513567] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463744, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.114016} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.513784] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 630.514625] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69fb93f1-d8bf-4187-8dbd-ed9a744ac45c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.539293] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] 1ef96606-d960-4283-b372-e2bad47f9367/1ef96606-d960-4283-b372-e2bad47f9367.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 630.539693] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9406491-7ed6-4442-b627-eed042184bd5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.564786] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 630.564786] env[61728]: value = "task-463746" [ 630.564786] env[61728]: _type = "Task" [ 630.564786] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.574611] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463746, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.608239] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463745, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.681494] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463743, 'name': CreateVM_Task, 'duration_secs': 0.547658} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.681772] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 630.682529] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.682693] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.683106] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 630.683377] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc54c660-683d-4ecc-9c4b-ed60c008f9da {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.688334] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 630.688334] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fce7ff-882d-64a9-ad38-2d33de3ce97a" [ 630.688334] env[61728]: _type = "Task" [ 630.688334] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.696917] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fce7ff-882d-64a9-ad38-2d33de3ce97a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.074326] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463746, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.112207] env[61728]: DEBUG oslo_vmware.api [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463745, 'name': PowerOnVM_Task, 'duration_secs': 0.988334} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.116406] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 631.116406] env[61728]: INFO nova.compute.manager [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Took 9.28 seconds to spawn the instance on the hypervisor. [ 631.116406] env[61728]: DEBUG nova.compute.manager [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 631.117672] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40528825-2f96-431a-aef5-3e4b870a26b3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.122163] env[61728]: DEBUG nova.network.neutron [req-5cbb4239-5118-4085-adee-61f2f913cbef req-0f33487f-2039-49e4-88a9-b4a48c53e100 service nova] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Updated VIF entry in instance network info cache for port a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 631.122163] env[61728]: DEBUG nova.network.neutron [req-5cbb4239-5118-4085-adee-61f2f913cbef req-0f33487f-2039-49e4-88a9-b4a48c53e100 service nova] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Updating instance_info_cache with network_info: [{"id": "a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7", "address": "fa:16:3e:6d:36:9d", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.216", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2de5ad6-df", "ovs_interfaceid": "a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.198978] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fce7ff-882d-64a9-ad38-2d33de3ce97a, 'name': SearchDatastore_Task, 'duration_secs': 0.009999} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.202955] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.202955] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 631.202955] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.202955] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.203138] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 631.203553] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-26015a12-49f6-48f0-b307-7ec26bd7f71a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.212216] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 631.212216] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 631.213200] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2922c9c-11d8-4dde-a4ea-07eded194a88 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.222387] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 631.222387] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb32bd-1b74-4d1c-d74f-1b388a0132c2" [ 631.222387] env[61728]: _type = "Task" [ 631.222387] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.233039] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb32bd-1b74-4d1c-d74f-1b388a0132c2, 'name': SearchDatastore_Task, 'duration_secs': 0.00892} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.233855] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a4a8b1d-0c1e-4921-a9da-7ed834928919 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.239817] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 631.239817] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d0d4f3-ffa0-521d-58dc-6a4812da07d4" [ 631.239817] env[61728]: _type = "Task" [ 631.239817] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.250340] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d0d4f3-ffa0-521d-58dc-6a4812da07d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.382916] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f65e40-234f-4331-9807-8b3458fa12a1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.390693] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bba33e1-6115-44d4-96d6-2a80720944fc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.426125] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f9edb8-d82f-4ab3-bcd6-2704b4a333ce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.436461] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae67532-7a00-4b3e-adb5-371997732be8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.450889] env[61728]: DEBUG nova.compute.provider_tree [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 631.575036] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463746, 'name': ReconfigVM_Task, 'duration_secs': 0.590598} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.575036] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Reconfigured VM instance instance-0000001f to attach disk [datastore1] 1ef96606-d960-4283-b372-e2bad47f9367/1ef96606-d960-4283-b372-e2bad47f9367.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 631.575512] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7415c9ea-432f-4ecd-bf81-ae04e557cb6e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.581602] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 631.581602] env[61728]: value = "task-463747" [ 631.581602] env[61728]: _type = "Task" [ 631.581602] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.589523] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463747, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.632189] env[61728]: DEBUG oslo_concurrency.lockutils [req-5cbb4239-5118-4085-adee-61f2f913cbef req-0f33487f-2039-49e4-88a9-b4a48c53e100 service nova] Releasing lock "refresh_cache-7a60fcd0-01b0-40bc-93e7-6f5ce03232d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.636331] env[61728]: INFO nova.compute.manager [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Took 47.66 seconds to build instance. [ 631.753469] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d0d4f3-ffa0-521d-58dc-6a4812da07d4, 'name': SearchDatastore_Task, 'duration_secs': 0.019602} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.753784] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.754176] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4/7a60fcd0-01b0-40bc-93e7-6f5ce03232d4.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 631.754351] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a1b87e9-c946-4ad9-9b6e-879c17bb02b3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.761095] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 631.761095] env[61728]: value = "task-463748" [ 631.761095] env[61728]: _type = "Task" [ 631.761095] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.772246] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463748, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.811155] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "223b1f93-d33e-4566-8c8e-75234a491814" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.811430] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "223b1f93-d33e-4566-8c8e-75234a491814" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.974404] env[61728]: ERROR nova.scheduler.client.report [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [req-b54c493d-1e9e-4f58-b614-8ae132f133f5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b54c493d-1e9e-4f58-b614-8ae132f133f5"}]} [ 631.991699] env[61728]: DEBUG nova.scheduler.client.report [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 632.011301] env[61728]: DEBUG nova.scheduler.client.report [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 632.012145] env[61728]: DEBUG nova.compute.provider_tree [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 632.026262] env[61728]: DEBUG nova.scheduler.client.report [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 632.047455] env[61728]: DEBUG nova.scheduler.client.report [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 632.094514] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463747, 'name': Rename_Task, 'duration_secs': 0.226364} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.094871] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 632.095398] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-47d61243-dc33-4920-9b74-eaa615727a13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.102166] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 632.102166] env[61728]: value = "task-463749" [ 632.102166] env[61728]: _type = "Task" [ 632.102166] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.113917] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463749, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.142857] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15d69b7f-8065-42c0-9cd5-a1bff9c75197 tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Lock "fbd521c8-9618-4b5c-839e-50d8631fa7b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.653s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.275345] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463748, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.608962] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d95f5f-51a2-4cdf-942d-98fa200d0dd2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.618178] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463749, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.621744] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd8e410-e6cd-4f81-9e2c-65619d745c45 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.654062] env[61728]: DEBUG nova.compute.manager [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 632.657750] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60efd42e-4b9b-449f-a145-15fa7b798a19 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.665859] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98528ff3-0d45-4835-a366-caef259a4a39 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.682092] env[61728]: DEBUG nova.compute.provider_tree [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.772937] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463748, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.558544} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.773512] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4/7a60fcd0-01b0-40bc-93e7-6f5ce03232d4.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 632.773742] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 632.774010] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-264e4fea-adb5-41df-a716-8c9b019115bb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.779886] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 632.779886] env[61728]: value = "task-463750" [ 632.779886] env[61728]: _type = "Task" [ 632.779886] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.787582] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463750, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.111903] env[61728]: DEBUG oslo_vmware.api [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463749, 'name': PowerOnVM_Task, 'duration_secs': 0.970576} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.112191] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 633.112406] env[61728]: INFO nova.compute.manager [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Took 8.50 seconds to spawn the instance on the hypervisor. [ 633.112592] env[61728]: DEBUG nova.compute.manager [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 633.113364] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79fda87a-5397-4b19-9f18-daea9670b8f8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.185063] env[61728]: DEBUG nova.scheduler.client.report [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.188950] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.252630] env[61728]: DEBUG nova.compute.manager [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 633.253576] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39073bcf-62bc-47a6-b520-9e2391587214 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.289217] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463750, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067118} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.289506] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 633.290280] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d245c25-292d-4f6d-a0e4-777a8388f2cf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.312924] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Reconfiguring VM instance instance-00000020 to attach disk [datastore1] 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4/7a60fcd0-01b0-40bc-93e7-6f5ce03232d4.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 633.313899] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9669145-d0f9-42f3-aa5f-000e8c0108ed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.334479] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 633.334479] env[61728]: value = "task-463751" [ 633.334479] env[61728]: _type = "Task" [ 633.334479] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.343812] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463751, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.640155] env[61728]: INFO nova.compute.manager [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Took 44.77 seconds to build instance. [ 633.693243] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.921s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.693243] env[61728]: DEBUG nova.compute.manager [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 633.698141] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.176s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.702227] env[61728]: INFO nova.compute.claims [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.764360] env[61728]: INFO nova.compute.manager [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] instance snapshotting [ 633.767821] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6573029c-fe09-467c-b437-b169a0799395 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.794026] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3379e3f-ab40-43e4-8dfc-547cd29d3fe0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.845481] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463751, 'name': ReconfigVM_Task, 'duration_secs': 0.380378} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.845997] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Reconfigured VM instance instance-00000020 to attach disk [datastore1] 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4/7a60fcd0-01b0-40bc-93e7-6f5ce03232d4.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 633.846632] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b520bd4c-4b34-4737-bb93-61dee4f57c8e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.853412] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 633.853412] env[61728]: value = "task-463752" [ 633.853412] env[61728]: _type = "Task" [ 633.853412] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.861894] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463752, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.142886] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49df4d37-8152-4a67-9c68-8a4c298a7e34 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "1ef96606-d960-4283-b372-e2bad47f9367" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.337s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.208042] env[61728]: DEBUG nova.compute.utils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.209972] env[61728]: DEBUG nova.compute.manager [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 634.210167] env[61728]: DEBUG nova.network.neutron [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 634.304010] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 634.304356] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-088b1d6b-4f3a-4f9e-8ac2-1b61734655fc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.313145] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 634.313145] env[61728]: value = "task-463753" [ 634.313145] env[61728]: _type = "Task" [ 634.313145] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.318234] env[61728]: DEBUG nova.policy [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '90cb8e83bbf74874ae3fda46f1822fd0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7f48655e35894d6daf7065aad4e84c2a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 634.325130] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463753, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.362937] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463752, 'name': Rename_Task, 'duration_secs': 0.137263} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.363222] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 634.363467] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67e722ee-b8f5-4916-9f94-ed328525435b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.369380] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 634.369380] env[61728]: value = "task-463754" [ 634.369380] env[61728]: _type = "Task" [ 634.369380] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.377150] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463754, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.650630] env[61728]: DEBUG nova.compute.manager [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 634.716931] env[61728]: DEBUG nova.compute.manager [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 634.823950] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463753, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.883525] env[61728]: DEBUG oslo_vmware.api [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463754, 'name': PowerOnVM_Task, 'duration_secs': 0.480773} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.883875] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 634.884138] env[61728]: INFO nova.compute.manager [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Took 7.53 seconds to spawn the instance on the hypervisor. [ 634.884367] env[61728]: DEBUG nova.compute.manager [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 634.885368] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b625ccb-8e73-41f5-8dff-7fd4f0aad15d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.177873] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.280365] env[61728]: DEBUG nova.network.neutron [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Successfully created port: 01099011-ff85-47eb-b05c-11566d3a6c0c {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 635.297277] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aadbd5eb-b224-4224-8496-f59f38bed09c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.304888] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eca53cd-4497-4377-8518-2f28374d6b6a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.337184] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef6346e-ce2e-451b-a3ac-9b5cbe7b9633 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.345053] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463753, 'name': CreateSnapshot_Task, 'duration_secs': 0.615186} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.347122] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 635.347881] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6ad282-a0d2-4cf7-b225-a7fad3bae120 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.351075] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2946ad05-d0e2-4636-b59c-48ed1b523c8b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.370907] env[61728]: DEBUG nova.compute.provider_tree [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.405676] env[61728]: INFO nova.compute.manager [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Took 43.45 seconds to build instance. [ 635.738225] env[61728]: DEBUG nova.compute.manager [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 635.772026] env[61728]: DEBUG nova.virt.hardware [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 635.772321] env[61728]: DEBUG nova.virt.hardware [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 635.772489] env[61728]: DEBUG nova.virt.hardware [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.772708] env[61728]: DEBUG nova.virt.hardware [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 635.772821] env[61728]: DEBUG nova.virt.hardware [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.773376] env[61728]: DEBUG nova.virt.hardware [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 635.773817] env[61728]: DEBUG nova.virt.hardware [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 635.774205] env[61728]: DEBUG nova.virt.hardware [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 635.774409] env[61728]: DEBUG nova.virt.hardware [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 635.774586] env[61728]: DEBUG nova.virt.hardware [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 635.774776] env[61728]: DEBUG nova.virt.hardware [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.776035] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb77b441-5bf8-4d8b-8ff7-3bd0722c103f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.786168] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64da9e32-0a07-4091-bd4f-c48f4cde267e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.881160] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 635.882112] env[61728]: DEBUG nova.scheduler.client.report [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 635.885368] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-36e03799-9975-4562-96f7-a7eefa7f2e44 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.895845] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 635.895845] env[61728]: value = "task-463755" [ 635.895845] env[61728]: _type = "Task" [ 635.895845] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.905934] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463755, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.909015] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3917ae9e-5973-482b-aba9-2a23f16aa67c tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Lock "7a60fcd0-01b0-40bc-93e7-6f5ce03232d4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.321s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.391194] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.693s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.392012] env[61728]: DEBUG nova.compute.manager [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 636.395587] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.509s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.397035] env[61728]: INFO nova.compute.claims [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 636.416354] env[61728]: DEBUG nova.compute.manager [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 636.420275] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463755, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.460158] env[61728]: DEBUG nova.compute.manager [req-ec787462-3dbc-4fb7-a640-20787b110f8d req-4bde607d-eb18-4975-b1e3-e7b1018f2e86 service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Received event network-changed-3196d005-b79e-4907-9b3c-59331e4d1e40 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 636.460380] env[61728]: DEBUG nova.compute.manager [req-ec787462-3dbc-4fb7-a640-20787b110f8d req-4bde607d-eb18-4975-b1e3-e7b1018f2e86 service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Refreshing instance network info cache due to event network-changed-3196d005-b79e-4907-9b3c-59331e4d1e40. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 636.460636] env[61728]: DEBUG oslo_concurrency.lockutils [req-ec787462-3dbc-4fb7-a640-20787b110f8d req-4bde607d-eb18-4975-b1e3-e7b1018f2e86 service nova] Acquiring lock "refresh_cache-fbd521c8-9618-4b5c-839e-50d8631fa7b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.460806] env[61728]: DEBUG oslo_concurrency.lockutils [req-ec787462-3dbc-4fb7-a640-20787b110f8d req-4bde607d-eb18-4975-b1e3-e7b1018f2e86 service nova] Acquired lock "refresh_cache-fbd521c8-9618-4b5c-839e-50d8631fa7b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.462744] env[61728]: DEBUG nova.network.neutron [req-ec787462-3dbc-4fb7-a640-20787b110f8d req-4bde607d-eb18-4975-b1e3-e7b1018f2e86 service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Refreshing network info cache for port 3196d005-b79e-4907-9b3c-59331e4d1e40 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 636.897582] env[61728]: DEBUG nova.compute.utils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.899357] env[61728]: DEBUG nova.compute.manager [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 636.899483] env[61728]: DEBUG nova.network.neutron [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 636.926688] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463755, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.945537] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.954643] env[61728]: DEBUG nova.policy [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '01f2593c9faf4c8dbccbcc9536a76a7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f22617f7e1f4598b01ebb980c61a1ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 637.341448] env[61728]: DEBUG nova.network.neutron [req-ec787462-3dbc-4fb7-a640-20787b110f8d req-4bde607d-eb18-4975-b1e3-e7b1018f2e86 service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Updated VIF entry in instance network info cache for port 3196d005-b79e-4907-9b3c-59331e4d1e40. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 637.341946] env[61728]: DEBUG nova.network.neutron [req-ec787462-3dbc-4fb7-a640-20787b110f8d req-4bde607d-eb18-4975-b1e3-e7b1018f2e86 service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Updating instance_info_cache with network_info: [{"id": "3196d005-b79e-4907-9b3c-59331e4d1e40", "address": "fa:16:3e:d1:5d:8a", "network": {"id": "527f7a39-789a-433b-9495-f13d2b52f0e0", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-189900763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99db3c843e8c4c3397cabba65ae510ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3196d005-b7", "ovs_interfaceid": "3196d005-b79e-4907-9b3c-59331e4d1e40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.372572] env[61728]: DEBUG nova.network.neutron [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Successfully created port: 0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 637.402775] env[61728]: DEBUG nova.compute.manager [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 637.420503] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463755, 'name': CloneVM_Task} progress is 95%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.846846] env[61728]: DEBUG oslo_concurrency.lockutils [req-ec787462-3dbc-4fb7-a640-20787b110f8d req-4bde607d-eb18-4975-b1e3-e7b1018f2e86 service nova] Releasing lock "refresh_cache-fbd521c8-9618-4b5c-839e-50d8631fa7b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.919291] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463755, 'name': CloneVM_Task, 'duration_secs': 1.937239} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.921841] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Created linked-clone VM from snapshot [ 637.923165] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6a5273-a585-4aa6-8fdd-f73d25f1f626 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.931147] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Uploading image 590a60b2-a8bc-44c6-8d0b-20d421bf3bb1 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 637.975112] env[61728]: DEBUG oslo_vmware.rw_handles [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 637.975112] env[61728]: value = "vm-122014" [ 637.975112] env[61728]: _type = "VirtualMachine" [ 637.975112] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 637.977944] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f7d07c7a-99ed-41db-999b-2655e316c878 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.987166] env[61728]: DEBUG oslo_vmware.rw_handles [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lease: (returnval){ [ 637.987166] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5263a217-abeb-63ba-6df7-688e02b15587" [ 637.987166] env[61728]: _type = "HttpNfcLease" [ 637.987166] env[61728]: } obtained for exporting VM: (result){ [ 637.987166] env[61728]: value = "vm-122014" [ 637.987166] env[61728]: _type = "VirtualMachine" [ 637.987166] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 637.987493] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the lease: (returnval){ [ 637.987493] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5263a217-abeb-63ba-6df7-688e02b15587" [ 637.987493] env[61728]: _type = "HttpNfcLease" [ 637.987493] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 637.989162] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d747e3-0050-4762-a3ca-430eb50ec3c5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.001683] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92473c40-daf4-4a90-b6ee-d503ee57aba9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.005486] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 638.005486] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5263a217-abeb-63ba-6df7-688e02b15587" [ 638.005486] env[61728]: _type = "HttpNfcLease" [ 638.005486] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 638.035582] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a04dab3-58ff-4de0-9437-657a4c6e7ee8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.045019] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89697426-793a-4048-b9d2-5b0249ea1033 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.053574] env[61728]: DEBUG nova.network.neutron [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Successfully updated port: 01099011-ff85-47eb-b05c-11566d3a6c0c {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 638.061133] env[61728]: DEBUG nova.compute.provider_tree [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.424241] env[61728]: DEBUG nova.compute.manager [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 638.452835] env[61728]: DEBUG nova.virt.hardware [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 638.453131] env[61728]: DEBUG nova.virt.hardware [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 638.453296] env[61728]: DEBUG nova.virt.hardware [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 638.453511] env[61728]: DEBUG nova.virt.hardware [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 638.453634] env[61728]: DEBUG nova.virt.hardware [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 638.453861] env[61728]: DEBUG nova.virt.hardware [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 638.454190] env[61728]: DEBUG nova.virt.hardware [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 638.454190] env[61728]: DEBUG nova.virt.hardware [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 638.454776] env[61728]: DEBUG nova.virt.hardware [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 638.454776] env[61728]: DEBUG nova.virt.hardware [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 638.454776] env[61728]: DEBUG nova.virt.hardware [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 638.455564] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f164eef5-6ea6-488f-a742-4539152a1e6d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.464244] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63319d97-16c5-4f05-b3be-1a1f3156f612 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.497573] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 638.497573] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5263a217-abeb-63ba-6df7-688e02b15587" [ 638.497573] env[61728]: _type = "HttpNfcLease" [ 638.497573] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 638.497573] env[61728]: DEBUG oslo_vmware.rw_handles [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 638.497573] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5263a217-abeb-63ba-6df7-688e02b15587" [ 638.497573] env[61728]: _type = "HttpNfcLease" [ 638.497573] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 638.498082] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40318ccf-5ba0-4565-aad8-06dbd82d1719 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.506222] env[61728]: DEBUG oslo_vmware.rw_handles [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529a3dc1-3314-54ef-648a-871f6364cd9f/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 638.506580] env[61728]: DEBUG oslo_vmware.rw_handles [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529a3dc1-3314-54ef-648a-871f6364cd9f/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 638.568170] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquiring lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.568318] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquired lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.568463] env[61728]: DEBUG nova.network.neutron [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 638.570102] env[61728]: DEBUG nova.scheduler.client.report [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.617406] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-92c1f275-6a51-4012-a5cc-0fdbf31e04b5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.913554] env[61728]: INFO nova.compute.manager [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Rebuilding instance [ 638.938321] env[61728]: DEBUG nova.network.neutron [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Successfully updated port: 0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 638.978294] env[61728]: DEBUG nova.compute.manager [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 638.979349] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756141a8-4d95-4d7c-b333-169e8e4ef0ce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.080744] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.685s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.081348] env[61728]: DEBUG nova.compute.manager [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 639.084079] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquiring lock "7a60fcd0-01b0-40bc-93e7-6f5ce03232d4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.084321] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Lock "7a60fcd0-01b0-40bc-93e7-6f5ce03232d4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.084537] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquiring lock "7a60fcd0-01b0-40bc-93e7-6f5ce03232d4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.084937] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Lock "7a60fcd0-01b0-40bc-93e7-6f5ce03232d4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.085186] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Lock "7a60fcd0-01b0-40bc-93e7-6f5ce03232d4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.087230] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.091s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.088592] env[61728]: INFO nova.compute.claims [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 639.096940] env[61728]: INFO nova.compute.manager [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Terminating instance [ 639.099377] env[61728]: DEBUG nova.compute.manager [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 639.099620] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 639.101933] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe082f85-ac58-4310-8599-f3a02645aedb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.111300] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 639.111682] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2845f4c-acdb-4983-a908-a3b7feb900de {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.118310] env[61728]: DEBUG oslo_vmware.api [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 639.118310] env[61728]: value = "task-463757" [ 639.118310] env[61728]: _type = "Task" [ 639.118310] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.123239] env[61728]: DEBUG nova.network.neutron [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.134489] env[61728]: DEBUG oslo_vmware.api [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463757, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.284299] env[61728]: DEBUG nova.network.neutron [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Updating instance_info_cache with network_info: [{"id": "01099011-ff85-47eb-b05c-11566d3a6c0c", "address": "fa:16:3e:a5:b6:18", "network": {"id": "02d4ac43-311d-4e89-ac44-d40b3790b873", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-928797810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f48655e35894d6daf7065aad4e84c2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01099011-ff", "ovs_interfaceid": "01099011-ff85-47eb-b05c-11566d3a6c0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.443380] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.443380] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquired lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.443380] env[61728]: DEBUG nova.network.neutron [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 639.454254] env[61728]: DEBUG nova.compute.manager [req-1f3dbc4c-af19-4e1f-ae10-8108604961ac req-67d3d60b-ef0a-4542-b02b-e68ffd0af6b3 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Received event network-vif-plugged-01099011-ff85-47eb-b05c-11566d3a6c0c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 639.454476] env[61728]: DEBUG oslo_concurrency.lockutils [req-1f3dbc4c-af19-4e1f-ae10-8108604961ac req-67d3d60b-ef0a-4542-b02b-e68ffd0af6b3 service nova] Acquiring lock "e3116a87-ce12-4ec8-b6fa-70ec2127ebb0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.454913] env[61728]: DEBUG oslo_concurrency.lockutils [req-1f3dbc4c-af19-4e1f-ae10-8108604961ac req-67d3d60b-ef0a-4542-b02b-e68ffd0af6b3 service nova] Lock "e3116a87-ce12-4ec8-b6fa-70ec2127ebb0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.457348] env[61728]: DEBUG oslo_concurrency.lockutils [req-1f3dbc4c-af19-4e1f-ae10-8108604961ac req-67d3d60b-ef0a-4542-b02b-e68ffd0af6b3 service nova] Lock "e3116a87-ce12-4ec8-b6fa-70ec2127ebb0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.457348] env[61728]: DEBUG nova.compute.manager [req-1f3dbc4c-af19-4e1f-ae10-8108604961ac req-67d3d60b-ef0a-4542-b02b-e68ffd0af6b3 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] No waiting events found dispatching network-vif-plugged-01099011-ff85-47eb-b05c-11566d3a6c0c {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 639.457348] env[61728]: WARNING nova.compute.manager [req-1f3dbc4c-af19-4e1f-ae10-8108604961ac req-67d3d60b-ef0a-4542-b02b-e68ffd0af6b3 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Received unexpected event network-vif-plugged-01099011-ff85-47eb-b05c-11566d3a6c0c for instance with vm_state building and task_state spawning. [ 639.457348] env[61728]: DEBUG nova.compute.manager [req-1f3dbc4c-af19-4e1f-ae10-8108604961ac req-67d3d60b-ef0a-4542-b02b-e68ffd0af6b3 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Received event network-changed-01099011-ff85-47eb-b05c-11566d3a6c0c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 639.457348] env[61728]: DEBUG nova.compute.manager [req-1f3dbc4c-af19-4e1f-ae10-8108604961ac req-67d3d60b-ef0a-4542-b02b-e68ffd0af6b3 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Refreshing instance network info cache due to event network-changed-01099011-ff85-47eb-b05c-11566d3a6c0c. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 639.457596] env[61728]: DEBUG oslo_concurrency.lockutils [req-1f3dbc4c-af19-4e1f-ae10-8108604961ac req-67d3d60b-ef0a-4542-b02b-e68ffd0af6b3 service nova] Acquiring lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.494283] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 639.494976] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a784e5be-5166-4b77-9694-0e3250eb58ac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.508287] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 639.508287] env[61728]: value = "task-463758" [ 639.508287] env[61728]: _type = "Task" [ 639.508287] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.518918] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463758, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.590049] env[61728]: DEBUG nova.compute.utils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 639.593234] env[61728]: DEBUG nova.compute.manager [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 639.593234] env[61728]: DEBUG nova.network.neutron [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 639.631411] env[61728]: DEBUG oslo_vmware.api [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463757, 'name': PowerOffVM_Task, 'duration_secs': 0.205774} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.631942] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 639.632066] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 639.632271] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a438aa78-f064-4c09-8866-04bf3a1e0d5b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.646327] env[61728]: DEBUG nova.policy [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd95d37807034ff7a5bccc5e125ca05c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a3e73722fe784ae7bd9865edd73fca16', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 639.705144] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 639.705383] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 639.705577] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Deleting the datastore file [datastore1] 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 639.705847] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a838031-859a-4717-a82e-3b81267b1f5b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.716023] env[61728]: DEBUG oslo_vmware.api [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for the task: (returnval){ [ 639.716023] env[61728]: value = "task-463760" [ 639.716023] env[61728]: _type = "Task" [ 639.716023] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.724395] env[61728]: DEBUG oslo_vmware.api [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463760, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.771309] env[61728]: DEBUG nova.compute.manager [req-a9d6b218-d1e2-4239-8599-c74f4e1efad5 req-52a35442-c0d8-4a76-a7f6-8c9e38c58e8f service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Received event network-vif-plugged-0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 639.772046] env[61728]: DEBUG oslo_concurrency.lockutils [req-a9d6b218-d1e2-4239-8599-c74f4e1efad5 req-52a35442-c0d8-4a76-a7f6-8c9e38c58e8f service nova] Acquiring lock "0fb1192e-99f1-4469-b196-60df7eab8185-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.772362] env[61728]: DEBUG oslo_concurrency.lockutils [req-a9d6b218-d1e2-4239-8599-c74f4e1efad5 req-52a35442-c0d8-4a76-a7f6-8c9e38c58e8f service nova] Lock "0fb1192e-99f1-4469-b196-60df7eab8185-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.772707] env[61728]: DEBUG oslo_concurrency.lockutils [req-a9d6b218-d1e2-4239-8599-c74f4e1efad5 req-52a35442-c0d8-4a76-a7f6-8c9e38c58e8f service nova] Lock "0fb1192e-99f1-4469-b196-60df7eab8185-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.773026] env[61728]: DEBUG nova.compute.manager [req-a9d6b218-d1e2-4239-8599-c74f4e1efad5 req-52a35442-c0d8-4a76-a7f6-8c9e38c58e8f service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] No waiting events found dispatching network-vif-plugged-0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 639.773592] env[61728]: WARNING nova.compute.manager [req-a9d6b218-d1e2-4239-8599-c74f4e1efad5 req-52a35442-c0d8-4a76-a7f6-8c9e38c58e8f service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Received unexpected event network-vif-plugged-0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 for instance with vm_state building and task_state spawning. [ 639.786763] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Releasing lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.788811] env[61728]: DEBUG nova.compute.manager [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Instance network_info: |[{"id": "01099011-ff85-47eb-b05c-11566d3a6c0c", "address": "fa:16:3e:a5:b6:18", "network": {"id": "02d4ac43-311d-4e89-ac44-d40b3790b873", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-928797810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f48655e35894d6daf7065aad4e84c2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01099011-ff", "ovs_interfaceid": "01099011-ff85-47eb-b05c-11566d3a6c0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 639.789353] env[61728]: DEBUG oslo_concurrency.lockutils [req-1f3dbc4c-af19-4e1f-ae10-8108604961ac req-67d3d60b-ef0a-4542-b02b-e68ffd0af6b3 service nova] Acquired lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.791348] env[61728]: DEBUG nova.network.neutron [req-1f3dbc4c-af19-4e1f-ae10-8108604961ac req-67d3d60b-ef0a-4542-b02b-e68ffd0af6b3 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Refreshing network info cache for port 01099011-ff85-47eb-b05c-11566d3a6c0c {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 639.792815] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:b6:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd7b5f1ef-d4b9-4ec3-b047-17e4cb349d25', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01099011-ff85-47eb-b05c-11566d3a6c0c', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 639.802511] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Creating folder: Project (7f48655e35894d6daf7065aad4e84c2a). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 639.804170] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0d42c72d-07b0-4da8-833d-8125d42ede95 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.816816] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Created folder: Project (7f48655e35894d6daf7065aad4e84c2a) in parent group-v121913. [ 639.817310] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Creating folder: Instances. Parent ref: group-v122015. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 639.817610] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2d79ddb0-7cfb-45ee-b2fa-a9a489446ac0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.826571] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Created folder: Instances in parent group-v122015. [ 639.826823] env[61728]: DEBUG oslo.service.loopingcall [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.828260] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 639.828260] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f65daf07-9d41-4215-a63c-8fb7f46b496b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.852122] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 639.852122] env[61728]: value = "task-463763" [ 639.852122] env[61728]: _type = "Task" [ 639.852122] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.863965] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463763, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.991091] env[61728]: DEBUG nova.network.neutron [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.996722] env[61728]: DEBUG nova.network.neutron [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Successfully created port: 16958edb-78a0-4245-b286-406e92c6e241 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 640.025216] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463758, 'name': PowerOffVM_Task, 'duration_secs': 0.278903} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.026942] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 640.027287] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 640.028194] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82274c60-05e5-43ad-96f3-88f5f7b3779b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.040450] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 640.040450] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f330bb62-3922-402d-9e76-fdea793a497a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.097680] env[61728]: DEBUG nova.compute.manager [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 640.115513] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 640.115779] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 640.115960] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleting the datastore file [datastore1] 1ef96606-d960-4283-b372-e2bad47f9367 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 640.117071] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9b9304ed-59cb-4ac1-b5b6-048a1a3f38ac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.124850] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 640.124850] env[61728]: value = "task-463765" [ 640.124850] env[61728]: _type = "Task" [ 640.124850] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.136310] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463765, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.225795] env[61728]: DEBUG oslo_vmware.api [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Task: {'id': task-463760, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.292336} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.233690] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 640.233906] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 640.234108] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 640.234292] env[61728]: INFO nova.compute.manager [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Took 1.13 seconds to destroy the instance on the hypervisor. [ 640.234544] env[61728]: DEBUG oslo.service.loopingcall [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 640.235264] env[61728]: DEBUG nova.compute.manager [-] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 640.235366] env[61728]: DEBUG nova.network.neutron [-] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 640.334507] env[61728]: DEBUG nova.network.neutron [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Updating instance_info_cache with network_info: [{"id": "0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3", "address": "fa:16:3e:f1:97:78", "network": {"id": "105779cb-8a8c-49ab-8ea4-00d3c7878a6b", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1826811954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f22617f7e1f4598b01ebb980c61a1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b046d2a-a7", "ovs_interfaceid": "0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.367583] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463763, 'name': CreateVM_Task, 'duration_secs': 0.361629} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.367801] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 640.368584] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.369122] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.369357] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 640.369695] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43d09b5e-e32c-4845-a572-b5f843290549 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.374450] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Waiting for the task: (returnval){ [ 640.374450] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5250e808-313c-ebd6-b9d8-6f77c31a4c28" [ 640.374450] env[61728]: _type = "Task" [ 640.374450] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.387507] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5250e808-313c-ebd6-b9d8-6f77c31a4c28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.640249] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463765, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.361088} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.640249] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 640.640249] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 640.640605] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 640.681236] env[61728]: DEBUG nova.network.neutron [req-1f3dbc4c-af19-4e1f-ae10-8108604961ac req-67d3d60b-ef0a-4542-b02b-e68ffd0af6b3 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Updated VIF entry in instance network info cache for port 01099011-ff85-47eb-b05c-11566d3a6c0c. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 640.681636] env[61728]: DEBUG nova.network.neutron [req-1f3dbc4c-af19-4e1f-ae10-8108604961ac req-67d3d60b-ef0a-4542-b02b-e68ffd0af6b3 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Updating instance_info_cache with network_info: [{"id": "01099011-ff85-47eb-b05c-11566d3a6c0c", "address": "fa:16:3e:a5:b6:18", "network": {"id": "02d4ac43-311d-4e89-ac44-d40b3790b873", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-928797810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f48655e35894d6daf7065aad4e84c2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01099011-ff", "ovs_interfaceid": "01099011-ff85-47eb-b05c-11566d3a6c0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.691827] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c61dbad4-c787-47ed-bf0c-c098e2f85914 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.706770] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d43440e2-24d6-45dc-8999-7888620a3fe1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.754335] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8a77ba9-1435-4642-91c1-76bce89ddcb1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.764991] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ea9037-2acc-487d-84b4-8c3a5b472480 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.781448] env[61728]: DEBUG nova.compute.provider_tree [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 640.838701] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Releasing lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.839718] env[61728]: DEBUG nova.compute.manager [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Instance network_info: |[{"id": "0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3", "address": "fa:16:3e:f1:97:78", "network": {"id": "105779cb-8a8c-49ab-8ea4-00d3c7878a6b", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1826811954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f22617f7e1f4598b01ebb980c61a1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b046d2a-a7", "ovs_interfaceid": "0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 640.840243] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:97:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '77aa121f-8fb6-42f3-aaea-43addfe449b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 640.849696] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Creating folder: Project (6f22617f7e1f4598b01ebb980c61a1ec). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 640.850396] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-76bf3503-045e-4af9-86de-3afab98ddfc5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.862196] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Created folder: Project (6f22617f7e1f4598b01ebb980c61a1ec) in parent group-v121913. [ 640.862513] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Creating folder: Instances. Parent ref: group-v122018. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 640.862832] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53e78674-95bd-4d36-b039-2ec4f07f855d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.873982] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Created folder: Instances in parent group-v122018. [ 640.874273] env[61728]: DEBUG oslo.service.loopingcall [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 640.874481] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 640.874747] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b485f725-afce-4ad8-9489-4ba0b8da4b8c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.899252] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5250e808-313c-ebd6-b9d8-6f77c31a4c28, 'name': SearchDatastore_Task, 'duration_secs': 0.01945} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.902922] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.902922] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 640.902922] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.902922] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.903088] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 640.903088] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 640.903088] env[61728]: value = "task-463768" [ 640.903088] env[61728]: _type = "Task" [ 640.903088] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.903088] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-df48f4c5-e868-4fdb-85cf-cbd8dd118c1d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.913030] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463768, 'name': CreateVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.914373] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 640.914602] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 640.915493] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd8bdbbf-7faa-44ce-8731-d8d7d66461e4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.920354] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Waiting for the task: (returnval){ [ 640.920354] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52be2b95-de95-a219-7307-da23923b9fd4" [ 640.920354] env[61728]: _type = "Task" [ 640.920354] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.930025] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52be2b95-de95-a219-7307-da23923b9fd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.111826] env[61728]: DEBUG nova.compute.manager [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 641.137998] env[61728]: DEBUG nova.virt.hardware [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 641.137998] env[61728]: DEBUG nova.virt.hardware [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 641.137998] env[61728]: DEBUG nova.virt.hardware [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 641.138442] env[61728]: DEBUG nova.virt.hardware [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 641.138442] env[61728]: DEBUG nova.virt.hardware [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 641.138442] env[61728]: DEBUG nova.virt.hardware [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 641.138442] env[61728]: DEBUG nova.virt.hardware [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 641.138442] env[61728]: DEBUG nova.virt.hardware [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 641.138759] env[61728]: DEBUG nova.virt.hardware [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 641.143964] env[61728]: DEBUG nova.virt.hardware [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 641.143964] env[61728]: DEBUG nova.virt.hardware [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 641.143964] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b450801f-dee4-448b-8e49-a9221d99b7f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.164528] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60417aa2-0b41-4bb5-8100-5584efba2c07 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.185839] env[61728]: DEBUG oslo_concurrency.lockutils [req-1f3dbc4c-af19-4e1f-ae10-8108604961ac req-67d3d60b-ef0a-4542-b02b-e68ffd0af6b3 service nova] Releasing lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.201987] env[61728]: DEBUG nova.network.neutron [-] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.324417] env[61728]: DEBUG nova.scheduler.client.report [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 63 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 641.324692] env[61728]: DEBUG nova.compute.provider_tree [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 63 to 64 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 641.325119] env[61728]: DEBUG nova.compute.provider_tree [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 641.421847] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463768, 'name': CreateVM_Task, 'duration_secs': 0.417572} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.425823] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 641.427659] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.427659] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.427659] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 641.428348] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b3074d1-963d-4c38-a0ad-fcffccbee9d3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.436897] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52be2b95-de95-a219-7307-da23923b9fd4, 'name': SearchDatastore_Task, 'duration_secs': 0.012484} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.438860] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c11fefb-ebf2-419c-959b-976be143b398 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.441442] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 641.441442] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ed56e9-8a02-a2da-6a5e-dc1d38e24acb" [ 641.441442] env[61728]: _type = "Task" [ 641.441442] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.448400] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Waiting for the task: (returnval){ [ 641.448400] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52de2933-0a22-cb1e-2fa8-84bc5b3dd53a" [ 641.448400] env[61728]: _type = "Task" [ 641.448400] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.455623] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ed56e9-8a02-a2da-6a5e-dc1d38e24acb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.461202] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52de2933-0a22-cb1e-2fa8-84bc5b3dd53a, 'name': SearchDatastore_Task, 'duration_secs': 0.011832} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.461469] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.461805] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] e3116a87-ce12-4ec8-b6fa-70ec2127ebb0/e3116a87-ce12-4ec8-b6fa-70ec2127ebb0.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 641.462012] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e8a9318d-2f92-4051-9685-ba364e698a26 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.468698] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Waiting for the task: (returnval){ [ 641.468698] env[61728]: value = "task-463769" [ 641.468698] env[61728]: _type = "Task" [ 641.468698] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.478798] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-463769, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.693345] env[61728]: DEBUG nova.virt.hardware [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 641.693658] env[61728]: DEBUG nova.virt.hardware [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 641.693860] env[61728]: DEBUG nova.virt.hardware [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 641.694926] env[61728]: DEBUG nova.virt.hardware [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 641.695194] env[61728]: DEBUG nova.virt.hardware [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 641.695368] env[61728]: DEBUG nova.virt.hardware [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 641.695755] env[61728]: DEBUG nova.virt.hardware [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 641.695980] env[61728]: DEBUG nova.virt.hardware [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 641.696247] env[61728]: DEBUG nova.virt.hardware [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 641.696463] env[61728]: DEBUG nova.virt.hardware [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 641.696671] env[61728]: DEBUG nova.virt.hardware [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 641.699566] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c945228-b307-4eb0-92a8-9b60da430980 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.706223] env[61728]: INFO nova.compute.manager [-] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Took 1.47 seconds to deallocate network for instance. [ 641.709309] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52718cc5-6e42-4955-adff-01b9b21fb333 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.730348] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:1c:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c492f5cc-7ae0-4cab-823c-0d5dd8c60b26', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9742f60c-5589-42f0-b3e4-a7d802b28b63', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 641.739362] env[61728]: DEBUG oslo.service.loopingcall [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 641.739787] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 641.740097] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-adf47a20-c155-4cd4-88fc-0be0447ddf67 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.766489] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 641.766489] env[61728]: value = "task-463770" [ 641.766489] env[61728]: _type = "Task" [ 641.766489] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.778237] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463770, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.814788] env[61728]: DEBUG nova.network.neutron [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Successfully updated port: 16958edb-78a0-4245-b286-406e92c6e241 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 641.831857] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.744s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.832146] env[61728]: DEBUG nova.compute.manager [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 641.835420] env[61728]: DEBUG oslo_concurrency.lockutils [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.939s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.835700] env[61728]: DEBUG nova.objects.instance [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lazy-loading 'resources' on Instance uuid adb58292-62cd-4142-ad66-db014bcbdc49 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 641.954681] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ed56e9-8a02-a2da-6a5e-dc1d38e24acb, 'name': SearchDatastore_Task, 'duration_secs': 0.016113} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.957790] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.957790] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 641.957790] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.957790] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.958152] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 641.958152] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3e307189-e798-4e73-930d-175d1173f57e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.970896] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 641.971109] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 641.974922] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac9b5bab-8ff3-4e10-984f-f9f928015d8d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.983757] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 641.983757] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528c6a6e-05c9-048a-933e-19fdc4e89f9e" [ 641.983757] env[61728]: _type = "Task" [ 641.983757] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.988519] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-463769, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.999876] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528c6a6e-05c9-048a-933e-19fdc4e89f9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.044258] env[61728]: DEBUG nova.compute.manager [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 642.045169] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6dc47f-f7c1-4ca6-8bcd-8a672bbb77b9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.219351] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.277737] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463770, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.318725] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Acquiring lock "refresh_cache-4fcc957e-5175-4589-b4da-d8451d92c815" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.318910] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Acquired lock "refresh_cache-4fcc957e-5175-4589-b4da-d8451d92c815" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.319105] env[61728]: DEBUG nova.network.neutron [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 642.337037] env[61728]: DEBUG nova.compute.utils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 642.338621] env[61728]: DEBUG nova.compute.manager [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 642.338748] env[61728]: DEBUG nova.network.neutron [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 642.409942] env[61728]: DEBUG nova.policy [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e5d859723326401a964d739e596a0257', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad9e831db0274681ad40c11a9bc6399e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 642.488564] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-463769, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.577852} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.493173] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] e3116a87-ce12-4ec8-b6fa-70ec2127ebb0/e3116a87-ce12-4ec8-b6fa-70ec2127ebb0.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 642.493173] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 642.493173] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0da74d47-ed44-4f50-a148-c2eea47901ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.503951] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528c6a6e-05c9-048a-933e-19fdc4e89f9e, 'name': SearchDatastore_Task, 'duration_secs': 0.045555} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.505946] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Waiting for the task: (returnval){ [ 642.505946] env[61728]: value = "task-463771" [ 642.505946] env[61728]: _type = "Task" [ 642.505946] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.506187] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-874b74c3-2423-4548-b154-9b8bdb1f2429 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.520354] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-463771, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.520749] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 642.520749] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522923e2-6e3e-3544-617b-6f124d726397" [ 642.520749] env[61728]: _type = "Task" [ 642.520749] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.536673] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522923e2-6e3e-3544-617b-6f124d726397, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.561718] env[61728]: INFO nova.compute.manager [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] instance snapshotting [ 642.562484] env[61728]: DEBUG nova.objects.instance [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'flavor' on Instance uuid 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 642.650878] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.654528] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.782695] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463770, 'name': CreateVM_Task, 'duration_secs': 0.565518} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.783022] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 642.783941] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.783941] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.784263] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 642.784788] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0f25820-a32f-4a95-a605-cf66d6c1f0ef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.790828] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 642.790828] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a06944-93f3-a4b8-5c02-2d5d6a4c54d2" [ 642.790828] env[61728]: _type = "Task" [ 642.790828] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.798354] env[61728]: DEBUG nova.network.neutron [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Successfully created port: aae5c541-394b-4d2d-ac32-f2550bb74e28 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 642.808324] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a06944-93f3-a4b8-5c02-2d5d6a4c54d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.829358] env[61728]: DEBUG nova.compute.manager [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Received event network-changed-0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 642.829747] env[61728]: DEBUG nova.compute.manager [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Refreshing instance network info cache due to event network-changed-0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 642.830180] env[61728]: DEBUG oslo_concurrency.lockutils [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] Acquiring lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.830423] env[61728]: DEBUG oslo_concurrency.lockutils [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] Acquired lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.833021] env[61728]: DEBUG nova.network.neutron [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Refreshing network info cache for port 0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 642.848293] env[61728]: DEBUG nova.compute.manager [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 642.873049] env[61728]: DEBUG nova.network.neutron [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.981171] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b741d364-a2f2-4e64-a879-d18b016cf09e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.988937] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69f29dc-0e64-4652-bb12-19fd54d00dae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.021613] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d750f8-e8bb-459f-a165-94933f8cae13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.032646] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-463771, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089978} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.035414] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 643.036444] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f827191-e3ea-4869-9bc2-046cc9869f79 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.040116] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d42605-9edd-46b5-915f-4457975e9083 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.047303] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522923e2-6e3e-3544-617b-6f124d726397, 'name': SearchDatastore_Task, 'duration_secs': 0.018181} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.050601] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.050910] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 0fb1192e-99f1-4469-b196-60df7eab8185/0fb1192e-99f1-4469-b196-60df7eab8185.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 643.051273] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0bb627ec-c5b8-483e-b75a-4e474600206e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.079719] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Reconfiguring VM instance instance-00000021 to attach disk [datastore1] e3116a87-ce12-4ec8-b6fa-70ec2127ebb0/e3116a87-ce12-4ec8-b6fa-70ec2127ebb0.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 643.082142] env[61728]: DEBUG nova.compute.provider_tree [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.084018] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b91e3ad6-0e8c-438d-bab2-ac99494ffd5a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.099496] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fbea5f-9fe6-4801-8f27-8e29832fe198 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.103969] env[61728]: DEBUG nova.scheduler.client.report [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 643.108783] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 643.108783] env[61728]: value = "task-463772" [ 643.108783] env[61728]: _type = "Task" [ 643.108783] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.127725] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Waiting for the task: (returnval){ [ 643.127725] env[61728]: value = "task-463773" [ 643.127725] env[61728]: _type = "Task" [ 643.127725] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.129213] env[61728]: DEBUG nova.network.neutron [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Updating instance_info_cache with network_info: [{"id": "16958edb-78a0-4245-b286-406e92c6e241", "address": "fa:16:3e:0d:c4:15", "network": {"id": "913135db-d861-45f9-b317-ebef5b83d967", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-388599888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3e73722fe784ae7bd9865edd73fca16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7874ee7f-20c7-4bd8-a750-ed489e9acc65", "external-id": "nsx-vlan-transportzone-753", "segmentation_id": 753, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16958edb-78", "ovs_interfaceid": "16958edb-78a0-4245-b286-406e92c6e241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.131273] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ac3750-31fb-4f6c-be89-ff7391b6e01e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.144456] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-463772, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.148319] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-463773, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.207501] env[61728]: DEBUG nova.compute.manager [req-cfcd80ff-d774-4105-bd1b-f7dab2e2f6a2 req-8970eeb9-a53e-4ee4-83d5-0734be98369c service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Received event network-changed-3196d005-b79e-4907-9b3c-59331e4d1e40 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 643.207501] env[61728]: DEBUG nova.compute.manager [req-cfcd80ff-d774-4105-bd1b-f7dab2e2f6a2 req-8970eeb9-a53e-4ee4-83d5-0734be98369c service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Refreshing instance network info cache due to event network-changed-3196d005-b79e-4907-9b3c-59331e4d1e40. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 643.207501] env[61728]: DEBUG oslo_concurrency.lockutils [req-cfcd80ff-d774-4105-bd1b-f7dab2e2f6a2 req-8970eeb9-a53e-4ee4-83d5-0734be98369c service nova] Acquiring lock "refresh_cache-fbd521c8-9618-4b5c-839e-50d8631fa7b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.207501] env[61728]: DEBUG oslo_concurrency.lockutils [req-cfcd80ff-d774-4105-bd1b-f7dab2e2f6a2 req-8970eeb9-a53e-4ee4-83d5-0734be98369c service nova] Acquired lock "refresh_cache-fbd521c8-9618-4b5c-839e-50d8631fa7b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.207501] env[61728]: DEBUG nova.network.neutron [req-cfcd80ff-d774-4105-bd1b-f7dab2e2f6a2 req-8970eeb9-a53e-4ee4-83d5-0734be98369c service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Refreshing network info cache for port 3196d005-b79e-4907-9b3c-59331e4d1e40 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 643.312770] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a06944-93f3-a4b8-5c02-2d5d6a4c54d2, 'name': SearchDatastore_Task, 'duration_secs': 0.015349} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.313144] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.315502] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 643.315502] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.315502] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.315502] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 643.315502] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d4cdf683-8412-4b89-96c2-361eb7a39e92 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.323968] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 643.324191] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 643.324965] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12b68d18-ad39-411a-818f-ab0b17ff19a3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.330925] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 643.330925] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5227f46d-14dc-3ca9-861b-c986b00b4d01" [ 643.330925] env[61728]: _type = "Task" [ 643.330925] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.341296] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5227f46d-14dc-3ca9-861b-c986b00b4d01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.561993] env[61728]: DEBUG nova.network.neutron [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Updated VIF entry in instance network info cache for port 0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 643.562371] env[61728]: DEBUG nova.network.neutron [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Updating instance_info_cache with network_info: [{"id": "0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3", "address": "fa:16:3e:f1:97:78", "network": {"id": "105779cb-8a8c-49ab-8ea4-00d3c7878a6b", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1826811954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f22617f7e1f4598b01ebb980c61a1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b046d2a-a7", "ovs_interfaceid": "0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.610377] env[61728]: DEBUG oslo_concurrency.lockutils [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.775s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.613173] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.727s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.615309] env[61728]: INFO nova.compute.claims [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 643.641685] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Releasing lock "refresh_cache-4fcc957e-5175-4589-b4da-d8451d92c815" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.642177] env[61728]: DEBUG nova.compute.manager [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Instance network_info: |[{"id": "16958edb-78a0-4245-b286-406e92c6e241", "address": "fa:16:3e:0d:c4:15", "network": {"id": "913135db-d861-45f9-b317-ebef5b83d967", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-388599888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3e73722fe784ae7bd9865edd73fca16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7874ee7f-20c7-4bd8-a750-ed489e9acc65", "external-id": "nsx-vlan-transportzone-753", "segmentation_id": 753, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16958edb-78", "ovs_interfaceid": "16958edb-78a0-4245-b286-406e92c6e241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 643.642647] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-463772, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.648642] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:c4:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7874ee7f-20c7-4bd8-a750-ed489e9acc65', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '16958edb-78a0-4245-b286-406e92c6e241', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 643.658567] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Creating folder: Project (a3e73722fe784ae7bd9865edd73fca16). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 643.659794] env[61728]: INFO nova.scheduler.client.report [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Deleted allocations for instance adb58292-62cd-4142-ad66-db014bcbdc49 [ 643.662963] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 643.662963] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a8fa2be-ebf8-40c0-8770-b8f01b31f9e6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.671380] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d7aeb92e-d453-4e7a-9fdf-e7c26e40021e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.674112] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-463773, 'name': ReconfigVM_Task, 'duration_secs': 0.382771} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.674877] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Reconfigured VM instance instance-00000021 to attach disk [datastore1] e3116a87-ce12-4ec8-b6fa-70ec2127ebb0/e3116a87-ce12-4ec8-b6fa-70ec2127ebb0.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 643.676488] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-858e07b4-60d7-49f4-8a2e-4a420093dd7d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.679536] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 643.679536] env[61728]: value = "task-463775" [ 643.679536] env[61728]: _type = "Task" [ 643.679536] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.684271] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Created folder: Project (a3e73722fe784ae7bd9865edd73fca16) in parent group-v121913. [ 643.684461] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Creating folder: Instances. Parent ref: group-v122022. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 643.685671] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7ebe015-85cc-401a-99e7-6df4aa44aaa0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.687318] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Waiting for the task: (returnval){ [ 643.687318] env[61728]: value = "task-463776" [ 643.687318] env[61728]: _type = "Task" [ 643.687318] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.690684] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463775, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.700568] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-463776, 'name': Rename_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.702579] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Created folder: Instances in parent group-v122022. [ 643.702824] env[61728]: DEBUG oslo.service.loopingcall [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.703844] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 643.703844] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa64ebf0-712e-45f7-8400-3d6015e819c0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.728445] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 643.728445] env[61728]: value = "task-463778" [ 643.728445] env[61728]: _type = "Task" [ 643.728445] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.739700] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463778, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.794791] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "6df7b619-8cc4-4dd9-8596-22dc83234a8e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.795110] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "6df7b619-8cc4-4dd9-8596-22dc83234a8e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.844420] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5227f46d-14dc-3ca9-861b-c986b00b4d01, 'name': SearchDatastore_Task, 'duration_secs': 0.015064} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.847043] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70fba498-c4d5-495f-94ef-40f4574fc60a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.850847] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 643.850847] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5233e247-c6e5-6c58-f63c-19ddacd4cdd4" [ 643.850847] env[61728]: _type = "Task" [ 643.850847] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.859122] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5233e247-c6e5-6c58-f63c-19ddacd4cdd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.864493] env[61728]: DEBUG nova.compute.manager [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 643.892616] env[61728]: DEBUG nova.virt.hardware [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 643.892916] env[61728]: DEBUG nova.virt.hardware [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 643.893201] env[61728]: DEBUG nova.virt.hardware [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 643.893350] env[61728]: DEBUG nova.virt.hardware [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 643.893544] env[61728]: DEBUG nova.virt.hardware [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 643.893720] env[61728]: DEBUG nova.virt.hardware [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 643.894066] env[61728]: DEBUG nova.virt.hardware [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 643.894284] env[61728]: DEBUG nova.virt.hardware [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 643.894502] env[61728]: DEBUG nova.virt.hardware [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 643.894688] env[61728]: DEBUG nova.virt.hardware [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 643.894955] env[61728]: DEBUG nova.virt.hardware [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 643.896008] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb42bedc-f466-4771-9879-bdc10cef5c2e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.904248] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c711e8c0-47fd-4438-9271-22401b8115e6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.949261] env[61728]: DEBUG nova.network.neutron [req-cfcd80ff-d774-4105-bd1b-f7dab2e2f6a2 req-8970eeb9-a53e-4ee4-83d5-0734be98369c service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Updated VIF entry in instance network info cache for port 3196d005-b79e-4907-9b3c-59331e4d1e40. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 643.949261] env[61728]: DEBUG nova.network.neutron [req-cfcd80ff-d774-4105-bd1b-f7dab2e2f6a2 req-8970eeb9-a53e-4ee4-83d5-0734be98369c service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Updating instance_info_cache with network_info: [{"id": "3196d005-b79e-4907-9b3c-59331e4d1e40", "address": "fa:16:3e:d1:5d:8a", "network": {"id": "527f7a39-789a-433b-9495-f13d2b52f0e0", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-189900763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99db3c843e8c4c3397cabba65ae510ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3196d005-b7", "ovs_interfaceid": "3196d005-b79e-4907-9b3c-59331e4d1e40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.065350] env[61728]: DEBUG oslo_concurrency.lockutils [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] Releasing lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.065711] env[61728]: DEBUG nova.compute.manager [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Received event network-vif-deleted-a2de5ad6-df39-40c4-b63c-deb7f9c4f0a7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 644.065958] env[61728]: DEBUG nova.compute.manager [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Received event network-vif-plugged-16958edb-78a0-4245-b286-406e92c6e241 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 644.066308] env[61728]: DEBUG oslo_concurrency.lockutils [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] Acquiring lock "4fcc957e-5175-4589-b4da-d8451d92c815-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.066476] env[61728]: DEBUG oslo_concurrency.lockutils [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] Lock "4fcc957e-5175-4589-b4da-d8451d92c815-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.066679] env[61728]: DEBUG oslo_concurrency.lockutils [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] Lock "4fcc957e-5175-4589-b4da-d8451d92c815-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.067246] env[61728]: DEBUG nova.compute.manager [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] No waiting events found dispatching network-vif-plugged-16958edb-78a0-4245-b286-406e92c6e241 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 644.067246] env[61728]: WARNING nova.compute.manager [req-0a3a0ae3-1402-42a7-a707-0ead2760b91b req-903aaa6c-1a54-45a5-a61d-dfd4312afdc4 service nova] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Received unexpected event network-vif-plugged-16958edb-78a0-4245-b286-406e92c6e241 for instance with vm_state building and task_state spawning. [ 644.124345] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-463772, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.611627} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.124644] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 0fb1192e-99f1-4469-b196-60df7eab8185/0fb1192e-99f1-4469-b196-60df7eab8185.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 644.124867] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 644.125149] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e1b6aeb3-5860-472b-903b-c0165510fee6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.133681] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 644.133681] env[61728]: value = "task-463779" [ 644.133681] env[61728]: _type = "Task" [ 644.133681] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.146324] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-463779, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.173784] env[61728]: DEBUG oslo_concurrency.lockutils [None req-37587cf9-4f95-4521-b6e0-13428af7ef68 tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "adb58292-62cd-4142-ad66-db014bcbdc49" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.202s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.191814] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463775, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.202560] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-463776, 'name': Rename_Task, 'duration_secs': 0.161175} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.202891] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 644.203465] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c9397ead-b03a-47ac-8a81-07661a27c426 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.212030] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Waiting for the task: (returnval){ [ 644.212030] env[61728]: value = "task-463780" [ 644.212030] env[61728]: _type = "Task" [ 644.212030] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.220550] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-463780, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.239041] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463778, 'name': CreateVM_Task, 'duration_secs': 0.344141} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.239230] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 644.240166] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.240286] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.240697] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 644.240974] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c98d94cc-794d-4012-ba6c-8738763cd646 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.247828] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Waiting for the task: (returnval){ [ 644.247828] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52838697-4fec-55bd-2c12-0656c2c89c90" [ 644.247828] env[61728]: _type = "Task" [ 644.247828] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.257046] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52838697-4fec-55bd-2c12-0656c2c89c90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.363958] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5233e247-c6e5-6c58-f63c-19ddacd4cdd4, 'name': SearchDatastore_Task, 'duration_secs': 0.013943} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.364273] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.364532] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 1ef96606-d960-4283-b372-e2bad47f9367/1ef96606-d960-4283-b372-e2bad47f9367.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 644.364828] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b2a0586-4e8d-47ac-b33a-5cdd6fc58948 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.373031] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 644.373031] env[61728]: value = "task-463781" [ 644.373031] env[61728]: _type = "Task" [ 644.373031] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.386373] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463781, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.451989] env[61728]: DEBUG oslo_concurrency.lockutils [req-cfcd80ff-d774-4105-bd1b-f7dab2e2f6a2 req-8970eeb9-a53e-4ee4-83d5-0734be98369c service nova] Releasing lock "refresh_cache-fbd521c8-9618-4b5c-839e-50d8631fa7b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.494596] env[61728]: DEBUG nova.network.neutron [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Successfully updated port: aae5c541-394b-4d2d-ac32-f2550bb74e28 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 644.627935] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.627935] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.648672] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-463779, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074437} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.649345] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 644.651091] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be5f485-c0c0-4d0e-979e-7371135a879a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.678397] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Reconfiguring VM instance instance-00000022 to attach disk [datastore1] 0fb1192e-99f1-4469-b196-60df7eab8185/0fb1192e-99f1-4469-b196-60df7eab8185.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 644.680878] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80896a4d-d623-4e81-9e1e-9fbe3fffa2bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.708456] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463775, 'name': CreateSnapshot_Task, 'duration_secs': 0.806116} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.710243] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 644.710663] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 644.710663] env[61728]: value = "task-463782" [ 644.710663] env[61728]: _type = "Task" [ 644.710663] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.711750] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aacfe17d-1828-4b7e-bb42-7215d037c6ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.747172] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-463782, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.747470] env[61728]: DEBUG oslo_vmware.api [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-463780, 'name': PowerOnVM_Task, 'duration_secs': 0.50629} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.751684] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 644.751684] env[61728]: INFO nova.compute.manager [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Took 9.01 seconds to spawn the instance on the hypervisor. [ 644.751870] env[61728]: DEBUG nova.compute.manager [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 644.758700] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21485e07-4ea5-4f1d-8ba7-c5be74e55078 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.781897] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52838697-4fec-55bd-2c12-0656c2c89c90, 'name': SearchDatastore_Task, 'duration_secs': 0.017064} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.783920] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.784847] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 644.784847] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.784847] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.784847] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 644.785851] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a86585b-2307-4de4-bc19-33276a57a737 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.807091] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 644.807393] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 644.808199] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8346b1ba-f6ba-4611-8b17-a684b1a93dba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.816701] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Waiting for the task: (returnval){ [ 644.816701] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52223cb6-e691-e230-fb27-15eb211e9a0c" [ 644.816701] env[61728]: _type = "Task" [ 644.816701] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.834021] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52223cb6-e691-e230-fb27-15eb211e9a0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.887499] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463781, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.997637] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Acquiring lock "refresh_cache-69d7af16-7708-4df6-acca-093b6c1df1ba" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.997788] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Acquired lock "refresh_cache-69d7af16-7708-4df6-acca-093b6c1df1ba" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.997940] env[61728]: DEBUG nova.network.neutron [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 645.087738] env[61728]: DEBUG nova.compute.manager [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Received event network-changed-16958edb-78a0-4245-b286-406e92c6e241 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 645.088029] env[61728]: DEBUG nova.compute.manager [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Refreshing instance network info cache due to event network-changed-16958edb-78a0-4245-b286-406e92c6e241. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 645.088174] env[61728]: DEBUG oslo_concurrency.lockutils [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] Acquiring lock "refresh_cache-4fcc957e-5175-4589-b4da-d8451d92c815" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.088353] env[61728]: DEBUG oslo_concurrency.lockutils [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] Acquired lock "refresh_cache-4fcc957e-5175-4589-b4da-d8451d92c815" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.088535] env[61728]: DEBUG nova.network.neutron [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Refreshing network info cache for port 16958edb-78a0-4245-b286-406e92c6e241 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 645.226952] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-463782, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.255301] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 645.255633] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4f59c650-85fa-45f1-aa57-ecae01db109e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.266512] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 645.266512] env[61728]: value = "task-463783" [ 645.266512] env[61728]: _type = "Task" [ 645.266512] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.278231] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463783, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.279993] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68820bb2-d879-40d7-8a09-404cafe991a0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.291049] env[61728]: INFO nova.compute.manager [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Took 46.86 seconds to build instance. [ 645.299149] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7af466e-6323-4905-a6b8-c7901207133f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.337926] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0114a1-3e74-4751-919e-966be98b0d36 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.352199] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9a6bfc-ce1a-4d1e-a631-89a01efa3a4a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.356949] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52223cb6-e691-e230-fb27-15eb211e9a0c, 'name': SearchDatastore_Task, 'duration_secs': 0.056559} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.358528] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf23fc5b-cb16-4d1e-95f2-589530eb4ccc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.369629] env[61728]: DEBUG nova.compute.provider_tree [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 645.372114] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Waiting for the task: (returnval){ [ 645.372114] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526a62f1-aa40-80db-c4a0-b5f742c6f7ea" [ 645.372114] env[61728]: _type = "Task" [ 645.372114] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.383570] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526a62f1-aa40-80db-c4a0-b5f742c6f7ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.391186] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463781, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587052} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.391186] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 1ef96606-d960-4283-b372-e2bad47f9367/1ef96606-d960-4283-b372-e2bad47f9367.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 645.391186] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 645.391186] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-261c46ac-48b4-473b-8620-d91f759fb845 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.399057] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 645.399057] env[61728]: value = "task-463784" [ 645.399057] env[61728]: _type = "Task" [ 645.399057] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.408618] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463784, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.472597] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.472974] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.473361] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.473462] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.473629] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.478348] env[61728]: INFO nova.compute.manager [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Terminating instance [ 645.480540] env[61728]: DEBUG nova.compute.manager [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 645.480764] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 645.481767] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e644cdaf-7c9e-4243-b3bd-48db6f7273cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.490235] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 645.490505] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c4f0a86e-4ec9-483b-946e-e05c4e01512f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.497224] env[61728]: DEBUG oslo_vmware.api [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 645.497224] env[61728]: value = "task-463785" [ 645.497224] env[61728]: _type = "Task" [ 645.497224] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.509464] env[61728]: DEBUG oslo_vmware.api [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463785, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.545206] env[61728]: DEBUG nova.network.neutron [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.726376] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-463782, 'name': ReconfigVM_Task, 'duration_secs': 0.578704} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.726674] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Reconfigured VM instance instance-00000022 to attach disk [datastore1] 0fb1192e-99f1-4469-b196-60df7eab8185/0fb1192e-99f1-4469-b196-60df7eab8185.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 645.727415] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6587bd9d-c775-4597-ba71-dd0c424088ac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.736355] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 645.736355] env[61728]: value = "task-463786" [ 645.736355] env[61728]: _type = "Task" [ 645.736355] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.752220] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-463786, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.778447] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463783, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.795522] env[61728]: DEBUG nova.network.neutron [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Updating instance_info_cache with network_info: [{"id": "aae5c541-394b-4d2d-ac32-f2550bb74e28", "address": "fa:16:3e:ae:32:7a", "network": {"id": "c6c9ab2b-a502-4a6d-bba2-05b649800f60", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-230929629-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad9e831db0274681ad40c11a9bc6399e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5bd281ed-ae39-485f-90ee-4ee27994b5b0", "external-id": "nsx-vlan-transportzone-305", "segmentation_id": 305, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaae5c541-39", "ovs_interfaceid": "aae5c541-394b-4d2d-ac32-f2550bb74e28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.800286] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4771677-83ff-4927-9b03-74c155a1f6c9 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Lock "e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.465s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.887808] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526a62f1-aa40-80db-c4a0-b5f742c6f7ea, 'name': SearchDatastore_Task, 'duration_secs': 0.016621} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.888133] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.888383] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 4fcc957e-5175-4589-b4da-d8451d92c815/4fcc957e-5175-4589-b4da-d8451d92c815.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 645.890035] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b57e00a2-304a-48af-b6b6-dc348d9820e6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.896831] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Waiting for the task: (returnval){ [ 645.896831] env[61728]: value = "task-463787" [ 645.896831] env[61728]: _type = "Task" [ 645.896831] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.904182] env[61728]: ERROR nova.scheduler.client.report [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [req-cbfaee05-04b0-4de8-b5e1-30d2178d09c3] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-cbfaee05-04b0-4de8-b5e1-30d2178d09c3"}]} [ 645.919547] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463784, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.138778} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.920102] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Acquiring lock "fbd521c8-9618-4b5c-839e-50d8631fa7b9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.920377] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Lock "fbd521c8-9618-4b5c-839e-50d8631fa7b9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.921104] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Acquiring lock "fbd521c8-9618-4b5c-839e-50d8631fa7b9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.921104] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Lock "fbd521c8-9618-4b5c-839e-50d8631fa7b9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.921104] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Lock "fbd521c8-9618-4b5c-839e-50d8631fa7b9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.922861] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463787, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.923956] env[61728]: DEBUG nova.scheduler.client.report [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 645.926631] env[61728]: DEBUG nova.network.neutron [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Updated VIF entry in instance network info cache for port 16958edb-78a0-4245-b286-406e92c6e241. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 645.926960] env[61728]: DEBUG nova.network.neutron [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Updating instance_info_cache with network_info: [{"id": "16958edb-78a0-4245-b286-406e92c6e241", "address": "fa:16:3e:0d:c4:15", "network": {"id": "913135db-d861-45f9-b317-ebef5b83d967", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-388599888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3e73722fe784ae7bd9865edd73fca16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7874ee7f-20c7-4bd8-a750-ed489e9acc65", "external-id": "nsx-vlan-transportzone-753", "segmentation_id": 753, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16958edb-78", "ovs_interfaceid": "16958edb-78a0-4245-b286-406e92c6e241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.928155] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 645.933053] env[61728]: INFO nova.compute.manager [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Terminating instance [ 645.937025] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23cd215-3275-4699-8f35-b98936143065 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.938640] env[61728]: DEBUG nova.compute.manager [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 645.938852] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 645.940435] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-750b7117-887f-4c10-a9c1-227c3e7a8102 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.945014] env[61728]: DEBUG nova.scheduler.client.report [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 645.945296] env[61728]: DEBUG nova.compute.provider_tree [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 645.967798] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] 1ef96606-d960-4283-b372-e2bad47f9367/1ef96606-d960-4283-b372-e2bad47f9367.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 645.968201] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 645.969186] env[61728]: DEBUG nova.scheduler.client.report [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 645.971490] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19a077b4-530b-4188-aae1-2b3a211598b7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.985798] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-83090ee9-ca0a-4e83-baf8-139870df00ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.994681] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 645.994681] env[61728]: value = "task-463789" [ 645.994681] env[61728]: _type = "Task" [ 645.994681] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.996171] env[61728]: DEBUG oslo_vmware.api [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Waiting for the task: (returnval){ [ 645.996171] env[61728]: value = "task-463788" [ 645.996171] env[61728]: _type = "Task" [ 645.996171] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.009168] env[61728]: DEBUG nova.scheduler.client.report [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 646.025899] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463789, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.026623] env[61728]: DEBUG oslo_vmware.api [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463785, 'name': PowerOffVM_Task, 'duration_secs': 0.254432} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.026879] env[61728]: DEBUG oslo_vmware.api [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463788, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.027147] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 646.027342] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 646.027675] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b564e9e5-cd13-4099-80fc-edf07aa25ae1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.112539] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 646.112946] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 646.113255] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Deleting the datastore file [datastore1] 26220c0c-f535-4315-b3bb-2e5ac4d2286e {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 646.113613] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b350bd88-19f1-4a01-a5c0-6cea8dae9318 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.124827] env[61728]: DEBUG oslo_vmware.api [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for the task: (returnval){ [ 646.124827] env[61728]: value = "task-463791" [ 646.124827] env[61728]: _type = "Task" [ 646.124827] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.135018] env[61728]: DEBUG oslo_vmware.api [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463791, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.250293] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-463786, 'name': Rename_Task, 'duration_secs': 0.465566} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.250604] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 646.250898] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98259eb3-d78a-4f7b-88c5-fe58efeb0e27 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.262870] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 646.262870] env[61728]: value = "task-463792" [ 646.262870] env[61728]: _type = "Task" [ 646.262870] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.276640] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-463792, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.285305] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463783, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.300262] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Releasing lock "refresh_cache-69d7af16-7708-4df6-acca-093b6c1df1ba" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.300486] env[61728]: DEBUG nova.compute.manager [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Instance network_info: |[{"id": "aae5c541-394b-4d2d-ac32-f2550bb74e28", "address": "fa:16:3e:ae:32:7a", "network": {"id": "c6c9ab2b-a502-4a6d-bba2-05b649800f60", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-230929629-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad9e831db0274681ad40c11a9bc6399e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5bd281ed-ae39-485f-90ee-4ee27994b5b0", "external-id": "nsx-vlan-transportzone-305", "segmentation_id": 305, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaae5c541-39", "ovs_interfaceid": "aae5c541-394b-4d2d-ac32-f2550bb74e28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 646.300994] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:32:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5bd281ed-ae39-485f-90ee-4ee27994b5b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aae5c541-394b-4d2d-ac32-f2550bb74e28', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 646.310105] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Creating folder: Project (ad9e831db0274681ad40c11a9bc6399e). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 646.313973] env[61728]: DEBUG nova.compute.manager [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.316802] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-54d90c4b-1f0b-412d-9c09-fbf3847b6162 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.331655] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Created folder: Project (ad9e831db0274681ad40c11a9bc6399e) in parent group-v121913. [ 646.331723] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Creating folder: Instances. Parent ref: group-v122027. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 646.335189] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ddf6c41-272c-4f4b-9249-39738b2ea891 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.347732] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Created folder: Instances in parent group-v122027. [ 646.348054] env[61728]: DEBUG oslo.service.loopingcall [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 646.351189] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 646.352276] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e00a41e1-cb6e-49f5-b341-22c772e47982 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.381495] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 646.381495] env[61728]: value = "task-463795" [ 646.381495] env[61728]: _type = "Task" [ 646.381495] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.396616] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463795, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.411615] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463787, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.435628] env[61728]: DEBUG oslo_concurrency.lockutils [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] Releasing lock "refresh_cache-4fcc957e-5175-4589-b4da-d8451d92c815" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.435962] env[61728]: DEBUG nova.compute.manager [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Received event network-vif-plugged-aae5c541-394b-4d2d-ac32-f2550bb74e28 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 646.436235] env[61728]: DEBUG oslo_concurrency.lockutils [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] Acquiring lock "69d7af16-7708-4df6-acca-093b6c1df1ba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.436460] env[61728]: DEBUG oslo_concurrency.lockutils [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] Lock "69d7af16-7708-4df6-acca-093b6c1df1ba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.436629] env[61728]: DEBUG oslo_concurrency.lockutils [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] Lock "69d7af16-7708-4df6-acca-093b6c1df1ba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.436867] env[61728]: DEBUG nova.compute.manager [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] No waiting events found dispatching network-vif-plugged-aae5c541-394b-4d2d-ac32-f2550bb74e28 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 646.437105] env[61728]: WARNING nova.compute.manager [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Received unexpected event network-vif-plugged-aae5c541-394b-4d2d-ac32-f2550bb74e28 for instance with vm_state building and task_state spawning. [ 646.437296] env[61728]: DEBUG nova.compute.manager [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Received event network-changed-aae5c541-394b-4d2d-ac32-f2550bb74e28 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 646.437459] env[61728]: DEBUG nova.compute.manager [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Refreshing instance network info cache due to event network-changed-aae5c541-394b-4d2d-ac32-f2550bb74e28. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 646.437644] env[61728]: DEBUG oslo_concurrency.lockutils [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] Acquiring lock "refresh_cache-69d7af16-7708-4df6-acca-093b6c1df1ba" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.437793] env[61728]: DEBUG oslo_concurrency.lockutils [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] Acquired lock "refresh_cache-69d7af16-7708-4df6-acca-093b6c1df1ba" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.438203] env[61728]: DEBUG nova.network.neutron [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Refreshing network info cache for port aae5c541-394b-4d2d-ac32-f2550bb74e28 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 646.511597] env[61728]: DEBUG oslo_vmware.api [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463788, 'name': PowerOffVM_Task, 'duration_secs': 0.221199} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.518486] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 646.518811] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 646.520137] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463789, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.520137] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d7adbef-7290-4f60-bd7c-88143f95c9ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.620682] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 646.621450] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 646.621450] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Deleting the datastore file [datastore1] fbd521c8-9618-4b5c-839e-50d8631fa7b9 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 646.621714] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85f84fbd-f89d-47bf-90d3-9ced3796e6ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.627062] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f390c3-2467-44f8-9fa9-ddc052391286 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.636456] env[61728]: DEBUG oslo_vmware.api [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Waiting for the task: (returnval){ [ 646.636456] env[61728]: value = "task-463797" [ 646.636456] env[61728]: _type = "Task" [ 646.636456] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.645479] env[61728]: DEBUG oslo_vmware.api [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463791, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.647285] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79368837-a5bc-48e0-bf42-abaa2df65d76 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.654106] env[61728]: DEBUG oslo_vmware.api [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463797, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.688431] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-149499ac-df52-4bac-80e0-0358e48ab632 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.697631] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ebe1d6-516f-47db-8458-850a2335c14c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.715831] env[61728]: DEBUG nova.compute.provider_tree [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 646.780633] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463783, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.786412] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-463792, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.841117] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.893261] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463795, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.908505] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463787, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.709973} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.909054] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 4fcc957e-5175-4589-b4da-d8451d92c815/4fcc957e-5175-4589-b4da-d8451d92c815.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 646.909200] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 646.909409] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5a83f6ee-8ef0-4e31-88f2-c79146a744f8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.916900] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Waiting for the task: (returnval){ [ 646.916900] env[61728]: value = "task-463798" [ 646.916900] env[61728]: _type = "Task" [ 646.916900] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.927835] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463798, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.008253] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463789, 'name': ReconfigVM_Task, 'duration_secs': 0.783433} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.008974] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Reconfigured VM instance instance-0000001f to attach disk [datastore1] 1ef96606-d960-4283-b372-e2bad47f9367/1ef96606-d960-4283-b372-e2bad47f9367.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 647.009799] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c3075f1-f086-4c65-8c9d-03875f854146 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.022065] env[61728]: DEBUG nova.compute.manager [req-c00acbbb-7935-427a-9178-77a7dc2927c8 req-51a276a8-a818-402a-ba7a-4a031836be31 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Received event network-changed-01099011-ff85-47eb-b05c-11566d3a6c0c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 647.022065] env[61728]: DEBUG nova.compute.manager [req-c00acbbb-7935-427a-9178-77a7dc2927c8 req-51a276a8-a818-402a-ba7a-4a031836be31 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Refreshing instance network info cache due to event network-changed-01099011-ff85-47eb-b05c-11566d3a6c0c. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 647.022065] env[61728]: DEBUG oslo_concurrency.lockutils [req-c00acbbb-7935-427a-9178-77a7dc2927c8 req-51a276a8-a818-402a-ba7a-4a031836be31 service nova] Acquiring lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.022065] env[61728]: DEBUG oslo_concurrency.lockutils [req-c00acbbb-7935-427a-9178-77a7dc2927c8 req-51a276a8-a818-402a-ba7a-4a031836be31 service nova] Acquired lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.022065] env[61728]: DEBUG nova.network.neutron [req-c00acbbb-7935-427a-9178-77a7dc2927c8 req-51a276a8-a818-402a-ba7a-4a031836be31 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Refreshing network info cache for port 01099011-ff85-47eb-b05c-11566d3a6c0c {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 647.026514] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 647.026514] env[61728]: value = "task-463799" [ 647.026514] env[61728]: _type = "Task" [ 647.026514] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.037588] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463799, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.138665] env[61728]: DEBUG oslo_vmware.api [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Task: {'id': task-463791, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.557287} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.141933] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 647.142218] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 647.142419] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 647.142794] env[61728]: INFO nova.compute.manager [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Took 1.66 seconds to destroy the instance on the hypervisor. [ 647.142912] env[61728]: DEBUG oslo.service.loopingcall [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 647.143832] env[61728]: DEBUG nova.compute.manager [-] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 647.143832] env[61728]: DEBUG nova.network.neutron [-] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 647.153524] env[61728]: DEBUG oslo_vmware.api [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Task: {'id': task-463797, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.276587} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.153896] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 647.154279] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 647.154611] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 647.154907] env[61728]: INFO nova.compute.manager [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Took 1.22 seconds to destroy the instance on the hypervisor. [ 647.155268] env[61728]: DEBUG oslo.service.loopingcall [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 647.155566] env[61728]: DEBUG nova.compute.manager [-] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 647.155784] env[61728]: DEBUG nova.network.neutron [-] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 647.257257] env[61728]: DEBUG nova.network.neutron [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Updated VIF entry in instance network info cache for port aae5c541-394b-4d2d-ac32-f2550bb74e28. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 647.257257] env[61728]: DEBUG nova.network.neutron [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Updating instance_info_cache with network_info: [{"id": "aae5c541-394b-4d2d-ac32-f2550bb74e28", "address": "fa:16:3e:ae:32:7a", "network": {"id": "c6c9ab2b-a502-4a6d-bba2-05b649800f60", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-230929629-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad9e831db0274681ad40c11a9bc6399e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5bd281ed-ae39-485f-90ee-4ee27994b5b0", "external-id": "nsx-vlan-transportzone-305", "segmentation_id": 305, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaae5c541-39", "ovs_interfaceid": "aae5c541-394b-4d2d-ac32-f2550bb74e28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.271200] env[61728]: DEBUG nova.scheduler.client.report [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 67 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 647.271930] env[61728]: DEBUG nova.compute.provider_tree [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 67 to 68 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 647.272264] env[61728]: DEBUG nova.compute.provider_tree [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 647.286605] env[61728]: DEBUG oslo_vmware.api [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-463792, 'name': PowerOnVM_Task, 'duration_secs': 0.840268} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.288992] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 647.289397] env[61728]: INFO nova.compute.manager [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Took 8.86 seconds to spawn the instance on the hypervisor. [ 647.289722] env[61728]: DEBUG nova.compute.manager [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 647.291655] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16949cf-5767-4252-b213-f0262ebff0ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.299127] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463783, 'name': CloneVM_Task, 'duration_secs': 1.91235} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.299883] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Created linked-clone VM from snapshot [ 647.300792] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-389cc70e-5049-4b3c-9f2f-4a9768ff780d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.316870] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Uploading image 4d578dcb-5948-4a97-a32f-c1442844b66f {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 647.357577] env[61728]: DEBUG oslo_vmware.rw_handles [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 647.357577] env[61728]: value = "vm-122026" [ 647.357577] env[61728]: _type = "VirtualMachine" [ 647.357577] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 647.357904] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5af90d1e-522f-4081-baa0-c8eeb1399b23 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.374678] env[61728]: DEBUG oslo_vmware.rw_handles [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lease: (returnval){ [ 647.374678] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5238cb04-7a96-b506-fbe1-5e64577c1bb1" [ 647.374678] env[61728]: _type = "HttpNfcLease" [ 647.374678] env[61728]: } obtained for exporting VM: (result){ [ 647.374678] env[61728]: value = "vm-122026" [ 647.374678] env[61728]: _type = "VirtualMachine" [ 647.374678] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 647.375010] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the lease: (returnval){ [ 647.375010] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5238cb04-7a96-b506-fbe1-5e64577c1bb1" [ 647.375010] env[61728]: _type = "HttpNfcLease" [ 647.375010] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 647.383910] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 647.383910] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5238cb04-7a96-b506-fbe1-5e64577c1bb1" [ 647.383910] env[61728]: _type = "HttpNfcLease" [ 647.383910] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 647.394879] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463795, 'name': CreateVM_Task, 'duration_secs': 0.622907} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.394879] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 647.395574] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.395738] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.396080] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 647.396339] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83c49b65-bab1-4753-8511-017b74c4341c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.402360] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Waiting for the task: (returnval){ [ 647.402360] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c44cd3-571d-97d8-23ea-159b8017e6da" [ 647.402360] env[61728]: _type = "Task" [ 647.402360] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.411230] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c44cd3-571d-97d8-23ea-159b8017e6da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.427708] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463798, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102187} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.427900] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 647.428929] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb56e2f3-e30c-465e-9e0e-e0a65daea2dd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.455112] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Reconfiguring VM instance instance-00000023 to attach disk [datastore1] 4fcc957e-5175-4589-b4da-d8451d92c815/4fcc957e-5175-4589-b4da-d8451d92c815.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 647.455472] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e08139f-4fd1-4b16-b687-58df9406f553 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.479071] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Waiting for the task: (returnval){ [ 647.479071] env[61728]: value = "task-463801" [ 647.479071] env[61728]: _type = "Task" [ 647.479071] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.487391] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463801, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.540278] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463799, 'name': Rename_Task, 'duration_secs': 0.256295} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.540620] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 647.540932] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e5d0612a-ddd4-426c-94ad-074df67cc71d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.548805] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 647.548805] env[61728]: value = "task-463802" [ 647.548805] env[61728]: _type = "Task" [ 647.548805] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.560549] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463802, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.731729] env[61728]: DEBUG nova.compute.manager [req-d7a52b0b-f545-4883-9d51-98f824a2247b req-4dced167-44cb-46ef-9e22-7c940062ab5b service nova] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Received event network-vif-deleted-89110ed6-c4a4-4f3b-b706-134ce068f320 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 647.732043] env[61728]: INFO nova.compute.manager [req-d7a52b0b-f545-4883-9d51-98f824a2247b req-4dced167-44cb-46ef-9e22-7c940062ab5b service nova] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Neutron deleted interface 89110ed6-c4a4-4f3b-b706-134ce068f320; detaching it from the instance and deleting it from the info cache [ 647.732363] env[61728]: DEBUG nova.network.neutron [req-d7a52b0b-f545-4883-9d51-98f824a2247b req-4dced167-44cb-46ef-9e22-7c940062ab5b service nova] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.759333] env[61728]: DEBUG oslo_concurrency.lockutils [req-9f627907-f9b8-4ddc-954f-4ab25f60c982 req-ae118271-536a-4bb3-aa60-da73709eb734 service nova] Releasing lock "refresh_cache-69d7af16-7708-4df6-acca-093b6c1df1ba" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.787168] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.174s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.787725] env[61728]: DEBUG nova.compute.manager [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 647.790774] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.698s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.791787] env[61728]: DEBUG nova.objects.instance [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lazy-loading 'resources' on Instance uuid db8e373d-e06c-43f0-a79c-2c1aaf65a01a {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 647.817089] env[61728]: INFO nova.compute.manager [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Took 46.31 seconds to build instance. [ 647.869361] env[61728]: DEBUG nova.network.neutron [req-c00acbbb-7935-427a-9178-77a7dc2927c8 req-51a276a8-a818-402a-ba7a-4a031836be31 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Updated VIF entry in instance network info cache for port 01099011-ff85-47eb-b05c-11566d3a6c0c. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 647.869361] env[61728]: DEBUG nova.network.neutron [req-c00acbbb-7935-427a-9178-77a7dc2927c8 req-51a276a8-a818-402a-ba7a-4a031836be31 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Updating instance_info_cache with network_info: [{"id": "01099011-ff85-47eb-b05c-11566d3a6c0c", "address": "fa:16:3e:a5:b6:18", "network": {"id": "02d4ac43-311d-4e89-ac44-d40b3790b873", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-928797810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f48655e35894d6daf7065aad4e84c2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01099011-ff", "ovs_interfaceid": "01099011-ff85-47eb-b05c-11566d3a6c0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.885823] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 647.885823] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5238cb04-7a96-b506-fbe1-5e64577c1bb1" [ 647.885823] env[61728]: _type = "HttpNfcLease" [ 647.885823] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 647.886901] env[61728]: DEBUG oslo_vmware.rw_handles [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 647.886901] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5238cb04-7a96-b506-fbe1-5e64577c1bb1" [ 647.886901] env[61728]: _type = "HttpNfcLease" [ 647.886901] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 647.887154] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2ccc23-6bb7-4aaf-8b38-1773952c7f81 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.897359] env[61728]: DEBUG oslo_vmware.rw_handles [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52db8589-979d-33f8-2e74-23b708bfebf8/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 647.897729] env[61728]: DEBUG oslo_vmware.rw_handles [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52db8589-979d-33f8-2e74-23b708bfebf8/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 647.968461] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c44cd3-571d-97d8-23ea-159b8017e6da, 'name': SearchDatastore_Task, 'duration_secs': 0.030178} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.968816] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.969132] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 647.969422] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.969641] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.969863] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 647.970178] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2d93c75-d780-4843-9791-5fdde908f465 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.982482] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 647.982736] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 647.983892] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddd3c026-cf2f-4df1-bbb0-3f67b8cdfa6d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.989954] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.994759] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Waiting for the task: (returnval){ [ 647.994759] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5287edc7-049a-faaf-88c9-c8dad5efc16b" [ 647.994759] env[61728]: _type = "Task" [ 647.994759] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.004374] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5287edc7-049a-faaf-88c9-c8dad5efc16b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.029736] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4e804f0a-e49d-4a46-bdef-845cb6aefdb4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.065317] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463802, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.144377] env[61728]: DEBUG nova.network.neutron [-] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.154980] env[61728]: DEBUG nova.network.neutron [-] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.236328] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b4e216a9-14d1-4c06-9fff-ee3fa478373e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.247300] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8b653b-fdc4-4203-abc6-596f67fda9d9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.291137] env[61728]: DEBUG nova.compute.manager [req-d7a52b0b-f545-4883-9d51-98f824a2247b req-4dced167-44cb-46ef-9e22-7c940062ab5b service nova] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Detach interface failed, port_id=89110ed6-c4a4-4f3b-b706-134ce068f320, reason: Instance 26220c0c-f535-4315-b3bb-2e5ac4d2286e could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 648.296137] env[61728]: DEBUG nova.compute.utils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.302591] env[61728]: DEBUG nova.compute.manager [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 648.302591] env[61728]: DEBUG nova.network.neutron [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 648.319956] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5dc6838-17bf-4c3a-b5d9-9d80e63700a6 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "0fb1192e-99f1-4469-b196-60df7eab8185" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.997s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.356306] env[61728]: DEBUG nova.policy [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05dd0234c7c8454ca9ba20cb5a7dfe25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '828ba999707e48f6bd9c11cfc74b1b2a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 648.368577] env[61728]: DEBUG oslo_concurrency.lockutils [req-c00acbbb-7935-427a-9178-77a7dc2927c8 req-51a276a8-a818-402a-ba7a-4a031836be31 service nova] Releasing lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.394292] env[61728]: DEBUG oslo_vmware.rw_handles [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529a3dc1-3314-54ef-648a-871f6364cd9f/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 648.395834] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27352bb2-638c-4843-808a-d00765aa1f59 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.406020] env[61728]: DEBUG oslo_vmware.rw_handles [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529a3dc1-3314-54ef-648a-871f6364cd9f/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 648.406020] env[61728]: ERROR oslo_vmware.rw_handles [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529a3dc1-3314-54ef-648a-871f6364cd9f/disk-0.vmdk due to incomplete transfer. [ 648.406020] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e182aa58-2d6c-42c4-8de2-02cbb1b5bfc7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.411348] env[61728]: DEBUG oslo_vmware.rw_handles [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529a3dc1-3314-54ef-648a-871f6364cd9f/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 648.411650] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Uploaded image 590a60b2-a8bc-44c6-8d0b-20d421bf3bb1 to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 648.414608] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 648.414976] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-458feff7-c635-4ea2-b9c8-b36b2ebb893a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.421092] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 648.421092] env[61728]: value = "task-463803" [ 648.421092] env[61728]: _type = "Task" [ 648.421092] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.439023] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463803, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.493241] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463801, 'name': ReconfigVM_Task, 'duration_secs': 0.608379} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.494243] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Reconfigured VM instance instance-00000023 to attach disk [datastore1] 4fcc957e-5175-4589-b4da-d8451d92c815/4fcc957e-5175-4589-b4da-d8451d92c815.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 648.495119] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3516f063-97a9-42ef-bb9f-6893bbc776d3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.510614] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5287edc7-049a-faaf-88c9-c8dad5efc16b, 'name': SearchDatastore_Task, 'duration_secs': 0.015068} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.513140] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Waiting for the task: (returnval){ [ 648.513140] env[61728]: value = "task-463804" [ 648.513140] env[61728]: _type = "Task" [ 648.513140] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.513140] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-585cb3b2-c501-4e40-b72b-acc03cb8de13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.527199] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Waiting for the task: (returnval){ [ 648.527199] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523c8f65-df68-d603-1407-fe8606b57cfd" [ 648.527199] env[61728]: _type = "Task" [ 648.527199] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.531429] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463804, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.542344] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523c8f65-df68-d603-1407-fe8606b57cfd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.561158] env[61728]: DEBUG oslo_vmware.api [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463802, 'name': PowerOnVM_Task, 'duration_secs': 0.782135} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.561277] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 648.561565] env[61728]: DEBUG nova.compute.manager [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 648.562796] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10bf932c-7e20-4528-ab8e-d5737e9fb36a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.648918] env[61728]: INFO nova.compute.manager [-] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Took 1.51 seconds to deallocate network for instance. [ 648.659180] env[61728]: INFO nova.compute.manager [-] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Took 1.50 seconds to deallocate network for instance. [ 648.749891] env[61728]: DEBUG nova.network.neutron [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Successfully created port: 8ee9bdfc-796a-498b-845f-02218a86801a {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 648.800633] env[61728]: DEBUG nova.compute.manager [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 648.825112] env[61728]: DEBUG nova.compute.manager [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 648.934196] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463803, 'name': Destroy_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.992149] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267a16d0-9b50-43e3-8cee-21bee8487759 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.000911] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd12cf3d-e000-4f02-bc77-0d9662e219b1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.040243] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec6b5dc-8397-4425-a2f2-0a73e9c110c8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.048317] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463804, 'name': Rename_Task, 'duration_secs': 0.234611} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.053702] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 649.054293] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523c8f65-df68-d603-1407-fe8606b57cfd, 'name': SearchDatastore_Task, 'duration_secs': 0.021066} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.054572] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6edeba7-4a0d-4577-8da6-564d5a5098f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.057431] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-554b9b30-1137-4bb6-949f-b9fd8304c9bd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.062900] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.063301] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 69d7af16-7708-4df6-acca-093b6c1df1ba/69d7af16-7708-4df6-acca-093b6c1df1ba.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 649.063635] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6e05e651-2356-45e6-9d4c-f01d421851ec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.082890] env[61728]: DEBUG nova.compute.provider_tree [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.088414] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Waiting for the task: (returnval){ [ 649.088414] env[61728]: value = "task-463805" [ 649.088414] env[61728]: _type = "Task" [ 649.088414] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.088730] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Waiting for the task: (returnval){ [ 649.088730] env[61728]: value = "task-463806" [ 649.088730] env[61728]: _type = "Task" [ 649.088730] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.089613] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.103405] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463805, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.106775] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463806, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.161639] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.170556] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.346671] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.432379] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463803, 'name': Destroy_Task, 'duration_secs': 0.537318} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.432706] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Destroyed the VM [ 649.433038] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 649.433348] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-518409b0-14ac-436d-9955-a667d3563f09 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.439855] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 649.439855] env[61728]: value = "task-463807" [ 649.439855] env[61728]: _type = "Task" [ 649.439855] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.447841] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463807, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.592917] env[61728]: DEBUG nova.scheduler.client.report [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.612253] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463805, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.616664] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463806, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.702071] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.702691] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.787920] env[61728]: DEBUG nova.compute.manager [req-1aa6bc38-1181-442e-abac-0556efdfc832 req-e6462eff-eca5-4269-9bb3-091092ff6b0e service nova] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Received event network-vif-deleted-3196d005-b79e-4907-9b3c-59331e4d1e40 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 649.810941] env[61728]: DEBUG nova.compute.manager [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 649.840338] env[61728]: DEBUG nova.virt.hardware [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 649.840687] env[61728]: DEBUG nova.virt.hardware [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 649.840933] env[61728]: DEBUG nova.virt.hardware [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 649.841234] env[61728]: DEBUG nova.virt.hardware [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 649.841469] env[61728]: DEBUG nova.virt.hardware [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 649.841813] env[61728]: DEBUG nova.virt.hardware [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 649.842246] env[61728]: DEBUG nova.virt.hardware [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 649.842557] env[61728]: DEBUG nova.virt.hardware [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 649.842945] env[61728]: DEBUG nova.virt.hardware [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 649.843266] env[61728]: DEBUG nova.virt.hardware [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 649.843546] env[61728]: DEBUG nova.virt.hardware [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 649.844910] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d7359e-c50f-4fb0-bd15-031709a03e6c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.857009] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da44e8c-efaf-4300-bb27-517b179f0a82 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.948948] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463807, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.100265] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.309s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.103306] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 27.440s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.119081] env[61728]: DEBUG oslo_vmware.api [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463805, 'name': PowerOnVM_Task, 'duration_secs': 0.742992} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.121839] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 650.122412] env[61728]: INFO nova.compute.manager [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Took 9.01 seconds to spawn the instance on the hypervisor. [ 650.122581] env[61728]: DEBUG nova.compute.manager [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 650.123514] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463806, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.124195] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec301bd6-d333-4755-9b2f-bf770a35f7a8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.127741] env[61728]: INFO nova.scheduler.client.report [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Deleted allocations for instance db8e373d-e06c-43f0-a79c-2c1aaf65a01a [ 650.322351] env[61728]: DEBUG nova.network.neutron [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Successfully updated port: 8ee9bdfc-796a-498b-845f-02218a86801a {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 650.450689] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463807, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.586523] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "1ef96606-d960-4283-b372-e2bad47f9367" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.586813] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "1ef96606-d960-4283-b372-e2bad47f9367" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.587053] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "1ef96606-d960-4283-b372-e2bad47f9367-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.587427] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "1ef96606-d960-4283-b372-e2bad47f9367-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.587723] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "1ef96606-d960-4283-b372-e2bad47f9367-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.590361] env[61728]: INFO nova.compute.manager [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Terminating instance [ 650.592501] env[61728]: DEBUG nova.compute.manager [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 650.592862] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 650.594565] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2203deb-64cf-439c-9671-0be7d2400e68 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.602840] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 650.602840] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e9cb93e-0f0f-46b2-bce5-1fb05f0f12b1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.612350] env[61728]: DEBUG oslo_vmware.api [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 650.612350] env[61728]: value = "task-463808" [ 650.612350] env[61728]: _type = "Task" [ 650.612350] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.628527] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463806, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.196884} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.633065] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 69d7af16-7708-4df6-acca-093b6c1df1ba/69d7af16-7708-4df6-acca-093b6c1df1ba.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 650.633392] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 650.637423] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f79e64a7-0e05-4914-a3b9-57a3ba1f351e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.640034] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0bfb9a05-1b68-45b7-9c2e-c30eee08cb23 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "db8e373d-e06c-43f0-a79c-2c1aaf65a01a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.732s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.654745] env[61728]: DEBUG oslo_vmware.api [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463808, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.659608] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Waiting for the task: (returnval){ [ 650.659608] env[61728]: value = "task-463809" [ 650.659608] env[61728]: _type = "Task" [ 650.659608] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.660416] env[61728]: INFO nova.compute.manager [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Took 44.79 seconds to build instance. [ 650.674607] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463809, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.824760] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Acquiring lock "refresh_cache-8c53c9c9-cdc1-407b-8827-8409cf137235" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.825111] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Acquired lock "refresh_cache-8c53c9c9-cdc1-407b-8827-8409cf137235" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.825364] env[61728]: DEBUG nova.network.neutron [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 650.950955] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463807, 'name': RemoveSnapshot_Task} progress is 31%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.127548] env[61728]: DEBUG oslo_vmware.api [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463808, 'name': PowerOffVM_Task, 'duration_secs': 0.297206} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.127841] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 651.128109] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 651.129740] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5b7320f1-965d-46d6-b370-2c31bd41687e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.141854] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "f2692e22-6c14-4bb3-a080-607f0731105d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.142095] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "f2692e22-6c14-4bb3-a080-607f0731105d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.170090] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 771341ed-8b8e-470c-9686-82650f5271b4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.170359] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 26220c0c-f535-4315-b3bb-2e5ac4d2286e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 651.170594] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance a3e9af38-5115-4337-a20e-5d6c6ec655fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.170779] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 620bb472-c36e-4c56-acdf-42e02614856b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.170966] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 8c4b7227-1a01-4400-beb8-2df27e17e329 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.171153] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance c8f1f4ae-ab08-4a03-a3fd-94014509e7e7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.171288] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.171404] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 2d389d5e-a53f-424e-90af-5ed05e2f93c7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.171517] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 4e5dc620-68a1-4e10-a8be-702c4999ca10 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.171679] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.171847] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance fc3c1d93-b095-4d80-8d71-243b66d85a46 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 651.171999] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 37b8ca5a-e0f8-414b-a363-e56db520f027 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.172142] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance f947d808-ebea-4593-b8e6-d3851c5216dc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.172257] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.172367] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 0f75629b-bdec-4593-aceb-cd0478972ccc actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.172490] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance fbd521c8-9618-4b5c-839e-50d8631fa7b9 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 651.172604] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 1ef96606-d960-4283-b372-e2bad47f9367 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.172727] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 651.172840] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance e3116a87-ce12-4ec8-b6fa-70ec2127ebb0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.172949] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 0fb1192e-99f1-4469-b196-60df7eab8185 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.173072] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 4fcc957e-5175-4589-b4da-d8451d92c815 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.173184] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 69d7af16-7708-4df6-acca-093b6c1df1ba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.173292] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 8c53c9c9-cdc1-407b-8827-8409cf137235 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.174808] env[61728]: DEBUG oslo_concurrency.lockutils [None req-72234733-b128-41a6-8000-724da897b407 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lock "4fcc957e-5175-4589-b4da-d8451d92c815" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.938s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.181846] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463809, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.106939} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.182120] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 651.183088] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2e600b-63e4-42e2-b26f-e2a2884cfeb8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.206300] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Reconfiguring VM instance instance-00000024 to attach disk [datastore1] 69d7af16-7708-4df6-acca-093b6c1df1ba/69d7af16-7708-4df6-acca-093b6c1df1ba.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 651.206920] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-821fce92-848e-4c7e-add5-c947e6f95018 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.226363] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Waiting for the task: (returnval){ [ 651.226363] env[61728]: value = "task-463811" [ 651.226363] env[61728]: _type = "Task" [ 651.226363] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.235021] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463811, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.369145] env[61728]: DEBUG nova.network.neutron [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.422455] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 651.422705] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 651.422908] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleting the datastore file [datastore1] 1ef96606-d960-4283-b372-e2bad47f9367 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 651.423215] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f92632d-d1ae-4e6e-afd6-af24ae6bb00c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.430806] env[61728]: DEBUG oslo_vmware.api [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 651.430806] env[61728]: value = "task-463812" [ 651.430806] env[61728]: _type = "Task" [ 651.430806] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.441730] env[61728]: DEBUG oslo_vmware.api [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463812, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.450169] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463807, 'name': RemoveSnapshot_Task} progress is 31%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.497025] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3e011526-06bf-464b-9af4-00c41cd153dd tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Acquiring lock "interface-4fcc957e-5175-4589-b4da-d8451d92c815-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.497376] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3e011526-06bf-464b-9af4-00c41cd153dd tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lock "interface-4fcc957e-5175-4589-b4da-d8451d92c815-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.497719] env[61728]: DEBUG nova.objects.instance [None req-3e011526-06bf-464b-9af4-00c41cd153dd tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lazy-loading 'flavor' on Instance uuid 4fcc957e-5175-4589-b4da-d8451d92c815 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 651.571698] env[61728]: DEBUG nova.network.neutron [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Updating instance_info_cache with network_info: [{"id": "8ee9bdfc-796a-498b-845f-02218a86801a", "address": "fa:16:3e:ac:f0:22", "network": {"id": "22454043-3114-444a-aed9-fe24700844e1", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1570154546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "828ba999707e48f6bd9c11cfc74b1b2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ee9bdfc-79", "ovs_interfaceid": "8ee9bdfc-796a-498b-845f-02218a86801a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.677316] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.679566] env[61728]: DEBUG nova.compute.manager [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 651.739649] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463811, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.941280] env[61728]: DEBUG oslo_vmware.api [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-463812, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.361448} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.941710] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 651.942076] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 651.942413] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 651.942741] env[61728]: INFO nova.compute.manager [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Took 1.35 seconds to destroy the instance on the hypervisor. [ 651.943144] env[61728]: DEBUG oslo.service.loopingcall [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 651.943534] env[61728]: DEBUG nova.compute.manager [-] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.943741] env[61728]: DEBUG nova.network.neutron [-] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 651.958810] env[61728]: DEBUG oslo_vmware.api [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463807, 'name': RemoveSnapshot_Task, 'duration_secs': 2.142388} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.959338] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 651.959720] env[61728]: INFO nova.compute.manager [None req-3b96b13c-956b-4f89-8c7a-1877755abece tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Took 18.19 seconds to snapshot the instance on the hypervisor. [ 652.003462] env[61728]: DEBUG nova.objects.instance [None req-3e011526-06bf-464b-9af4-00c41cd153dd tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lazy-loading 'pci_requests' on Instance uuid 4fcc957e-5175-4589-b4da-d8451d92c815 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 652.063238] env[61728]: DEBUG nova.compute.manager [req-7c371e87-9240-4a95-8dbe-5777abc10a46 req-32f10514-8a14-435d-a822-3cc131ad97d1 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Received event network-vif-plugged-8ee9bdfc-796a-498b-845f-02218a86801a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 652.063472] env[61728]: DEBUG oslo_concurrency.lockutils [req-7c371e87-9240-4a95-8dbe-5777abc10a46 req-32f10514-8a14-435d-a822-3cc131ad97d1 service nova] Acquiring lock "8c53c9c9-cdc1-407b-8827-8409cf137235-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.063844] env[61728]: DEBUG oslo_concurrency.lockutils [req-7c371e87-9240-4a95-8dbe-5777abc10a46 req-32f10514-8a14-435d-a822-3cc131ad97d1 service nova] Lock "8c53c9c9-cdc1-407b-8827-8409cf137235-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.063844] env[61728]: DEBUG oslo_concurrency.lockutils [req-7c371e87-9240-4a95-8dbe-5777abc10a46 req-32f10514-8a14-435d-a822-3cc131ad97d1 service nova] Lock "8c53c9c9-cdc1-407b-8827-8409cf137235-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.064099] env[61728]: DEBUG nova.compute.manager [req-7c371e87-9240-4a95-8dbe-5777abc10a46 req-32f10514-8a14-435d-a822-3cc131ad97d1 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] No waiting events found dispatching network-vif-plugged-8ee9bdfc-796a-498b-845f-02218a86801a {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 652.064195] env[61728]: WARNING nova.compute.manager [req-7c371e87-9240-4a95-8dbe-5777abc10a46 req-32f10514-8a14-435d-a822-3cc131ad97d1 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Received unexpected event network-vif-plugged-8ee9bdfc-796a-498b-845f-02218a86801a for instance with vm_state building and task_state spawning. [ 652.064353] env[61728]: DEBUG nova.compute.manager [req-7c371e87-9240-4a95-8dbe-5777abc10a46 req-32f10514-8a14-435d-a822-3cc131ad97d1 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Received event network-changed-8ee9bdfc-796a-498b-845f-02218a86801a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 652.064504] env[61728]: DEBUG nova.compute.manager [req-7c371e87-9240-4a95-8dbe-5777abc10a46 req-32f10514-8a14-435d-a822-3cc131ad97d1 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Refreshing instance network info cache due to event network-changed-8ee9bdfc-796a-498b-845f-02218a86801a. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 652.064674] env[61728]: DEBUG oslo_concurrency.lockutils [req-7c371e87-9240-4a95-8dbe-5777abc10a46 req-32f10514-8a14-435d-a822-3cc131ad97d1 service nova] Acquiring lock "refresh_cache-8c53c9c9-cdc1-407b-8827-8409cf137235" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.075148] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Releasing lock "refresh_cache-8c53c9c9-cdc1-407b-8827-8409cf137235" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.075461] env[61728]: DEBUG nova.compute.manager [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Instance network_info: |[{"id": "8ee9bdfc-796a-498b-845f-02218a86801a", "address": "fa:16:3e:ac:f0:22", "network": {"id": "22454043-3114-444a-aed9-fe24700844e1", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1570154546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "828ba999707e48f6bd9c11cfc74b1b2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ee9bdfc-79", "ovs_interfaceid": "8ee9bdfc-796a-498b-845f-02218a86801a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 652.076059] env[61728]: DEBUG oslo_concurrency.lockutils [req-7c371e87-9240-4a95-8dbe-5777abc10a46 req-32f10514-8a14-435d-a822-3cc131ad97d1 service nova] Acquired lock "refresh_cache-8c53c9c9-cdc1-407b-8827-8409cf137235" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.076257] env[61728]: DEBUG nova.network.neutron [req-7c371e87-9240-4a95-8dbe-5777abc10a46 req-32f10514-8a14-435d-a822-3cc131ad97d1 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Refreshing network info cache for port 8ee9bdfc-796a-498b-845f-02218a86801a {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 652.077766] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:f0:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0746f464-a938-427b-ba02-600449df5070', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ee9bdfc-796a-498b-845f-02218a86801a', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 652.088837] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Creating folder: Project (828ba999707e48f6bd9c11cfc74b1b2a). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 652.092898] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5709fb70-d0a4-4018-9e66-699121079bde {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.105379] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Created folder: Project (828ba999707e48f6bd9c11cfc74b1b2a) in parent group-v121913. [ 652.105581] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Creating folder: Instances. Parent ref: group-v122030. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 652.105825] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-079cb1aa-0891-4eb0-a2d8-f7d225966693 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.114561] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Created folder: Instances in parent group-v122030. [ 652.114832] env[61728]: DEBUG oslo.service.loopingcall [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 652.114997] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 652.115223] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e3974d6-c24f-40ad-bbbb-ae96e3a13ec6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.138280] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 652.138280] env[61728]: value = "task-463815" [ 652.138280] env[61728]: _type = "Task" [ 652.138280] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.147499] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463815, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.184404] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 507b5aba-aec9-438b-92c0-5efa17959bc7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.218153] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.237224] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463811, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.377613] env[61728]: DEBUG nova.network.neutron [req-7c371e87-9240-4a95-8dbe-5777abc10a46 req-32f10514-8a14-435d-a822-3cc131ad97d1 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Updated VIF entry in instance network info cache for port 8ee9bdfc-796a-498b-845f-02218a86801a. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 652.378010] env[61728]: DEBUG nova.network.neutron [req-7c371e87-9240-4a95-8dbe-5777abc10a46 req-32f10514-8a14-435d-a822-3cc131ad97d1 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Updating instance_info_cache with network_info: [{"id": "8ee9bdfc-796a-498b-845f-02218a86801a", "address": "fa:16:3e:ac:f0:22", "network": {"id": "22454043-3114-444a-aed9-fe24700844e1", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1570154546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "828ba999707e48f6bd9c11cfc74b1b2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ee9bdfc-79", "ovs_interfaceid": "8ee9bdfc-796a-498b-845f-02218a86801a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.508096] env[61728]: DEBUG nova.objects.base [None req-3e011526-06bf-464b-9af4-00c41cd153dd tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Object Instance<4fcc957e-5175-4589-b4da-d8451d92c815> lazy-loaded attributes: flavor,pci_requests {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 652.508392] env[61728]: DEBUG nova.network.neutron [None req-3e011526-06bf-464b-9af4-00c41cd153dd tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 652.592255] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3e011526-06bf-464b-9af4-00c41cd153dd tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lock "interface-4fcc957e-5175-4589-b4da-d8451d92c815-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.095s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.652617] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463815, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.689867] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 650d2875-68f7-4131-b9b7-9799ee08f1d6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.739216] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463811, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.819153] env[61728]: DEBUG nova.network.neutron [-] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.881806] env[61728]: DEBUG oslo_concurrency.lockutils [req-7c371e87-9240-4a95-8dbe-5777abc10a46 req-32f10514-8a14-435d-a822-3cc131ad97d1 service nova] Releasing lock "refresh_cache-8c53c9c9-cdc1-407b-8827-8409cf137235" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.150098] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463815, 'name': CreateVM_Task, 'duration_secs': 0.587529} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.150444] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 653.150979] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.151168] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.151491] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 653.151784] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f34ad3e6-9365-4496-8ace-4a048fe85db7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.156543] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Waiting for the task: (returnval){ [ 653.156543] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521608e0-e8c9-c0cf-b01f-08483ce3f538" [ 653.156543] env[61728]: _type = "Task" [ 653.156543] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.167037] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521608e0-e8c9-c0cf-b01f-08483ce3f538, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.194030] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance ad304c11-4ff8-4910-92a5-24216d727c4b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.237247] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463811, 'name': ReconfigVM_Task, 'duration_secs': 1.60373} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.237478] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Reconfigured VM instance instance-00000024 to attach disk [datastore1] 69d7af16-7708-4df6-acca-093b6c1df1ba/69d7af16-7708-4df6-acca-093b6c1df1ba.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 653.238134] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be2f1ed3-dc25-42b6-85a0-f65108d31bd5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.244812] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Waiting for the task: (returnval){ [ 653.244812] env[61728]: value = "task-463816" [ 653.244812] env[61728]: _type = "Task" [ 653.244812] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.252554] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463816, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.322135] env[61728]: INFO nova.compute.manager [-] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Took 1.38 seconds to deallocate network for instance. [ 653.668046] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521608e0-e8c9-c0cf-b01f-08483ce3f538, 'name': SearchDatastore_Task, 'duration_secs': 0.028723} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.668364] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.668603] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 653.668849] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.669009] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.669211] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 653.669494] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-77600e66-be6d-4bf1-a041-dd5fa1bd5852 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.691677] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 653.691996] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 653.692844] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29ee2339-c7f1-444d-9a81-6433d64dac13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.696121] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance b833f691-e76e-4a2e-94a4-7594fadc3ba8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.699161] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Waiting for the task: (returnval){ [ 653.699161] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c4dff3-0f37-94a4-10e5-12d07274ec39" [ 653.699161] env[61728]: _type = "Task" [ 653.699161] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.707228] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c4dff3-0f37-94a4-10e5-12d07274ec39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.755046] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463816, 'name': Rename_Task, 'duration_secs': 0.430235} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.755046] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 653.755046] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e2584cb-876f-4f57-bd5b-bd935551cac2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.760936] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Waiting for the task: (returnval){ [ 653.760936] env[61728]: value = "task-463817" [ 653.760936] env[61728]: _type = "Task" [ 653.760936] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.769518] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463817, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.832537] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.075083] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Acquiring lock "f947d808-ebea-4593-b8e6-d3851c5216dc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.075446] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lock "f947d808-ebea-4593-b8e6-d3851c5216dc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.075686] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Acquiring lock "f947d808-ebea-4593-b8e6-d3851c5216dc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.075932] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lock "f947d808-ebea-4593-b8e6-d3851c5216dc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.076170] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lock "f947d808-ebea-4593-b8e6-d3851c5216dc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.078828] env[61728]: INFO nova.compute.manager [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Terminating instance [ 654.080770] env[61728]: DEBUG nova.compute.manager [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 654.080976] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 654.081950] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4baaf382-87e3-42ea-bac9-998e697e2299 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.089970] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 654.090234] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-16536e89-c090-49ff-9aa2-41cc756f2533 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.096534] env[61728]: DEBUG oslo_vmware.api [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 654.096534] env[61728]: value = "task-463818" [ 654.096534] env[61728]: _type = "Task" [ 654.096534] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.104349] env[61728]: DEBUG oslo_vmware.api [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463818, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.201305] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 1d9114cb-a0af-4874-962f-27237b3c89cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.212050] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c4dff3-0f37-94a4-10e5-12d07274ec39, 'name': SearchDatastore_Task, 'duration_secs': 0.03534} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.212050] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac5536e9-4260-4faf-965d-818f2b2aa4ed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.217269] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Waiting for the task: (returnval){ [ 654.217269] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520e3307-8385-0e04-d0f1-924ba57ac86e" [ 654.217269] env[61728]: _type = "Task" [ 654.217269] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.228038] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520e3307-8385-0e04-d0f1-924ba57ac86e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.262475] env[61728]: DEBUG nova.compute.manager [req-8e02f103-dd0b-48d5-b120-b21bc404a4e1 req-622785ea-1e9b-4e3f-9937-a48362d5abc2 service nova] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Received event network-vif-deleted-9742f60c-5589-42f0-b3e4-a7d802b28b63 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 654.273231] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463817, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.606325] env[61728]: DEBUG oslo_vmware.api [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463818, 'name': PowerOffVM_Task, 'duration_secs': 0.287368} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.606667] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 654.606885] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 654.607189] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5888cfc-c643-4a3b-913f-c822760bbb50 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.668963] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 654.669248] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 654.669469] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Deleting the datastore file [datastore1] f947d808-ebea-4593-b8e6-d3851c5216dc {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 654.669820] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12ef55b4-3ec9-41d1-895b-3ac31bc12c8a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.676669] env[61728]: DEBUG oslo_vmware.api [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for the task: (returnval){ [ 654.676669] env[61728]: value = "task-463820" [ 654.676669] env[61728]: _type = "Task" [ 654.676669] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.685316] env[61728]: DEBUG oslo_vmware.api [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463820, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.706448] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance ca8d41dc-5719-4ade-b82a-b1d7f1c8def1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.730412] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520e3307-8385-0e04-d0f1-924ba57ac86e, 'name': SearchDatastore_Task, 'duration_secs': 0.020515} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.730696] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.730978] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 8c53c9c9-cdc1-407b-8827-8409cf137235/8c53c9c9-cdc1-407b-8827-8409cf137235.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 654.731296] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e459c1e-c9cd-45de-a650-24e05dd21ddb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.739599] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Waiting for the task: (returnval){ [ 654.739599] env[61728]: value = "task-463821" [ 654.739599] env[61728]: _type = "Task" [ 654.739599] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.744202] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Acquiring lock "4fcc957e-5175-4589-b4da-d8451d92c815" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.744262] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lock "4fcc957e-5175-4589-b4da-d8451d92c815" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.744449] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Acquiring lock "4fcc957e-5175-4589-b4da-d8451d92c815-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.744686] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lock "4fcc957e-5175-4589-b4da-d8451d92c815-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.744909] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lock "4fcc957e-5175-4589-b4da-d8451d92c815-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.748088] env[61728]: INFO nova.compute.manager [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Terminating instance [ 654.753913] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463821, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.754649] env[61728]: DEBUG nova.compute.manager [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 654.754946] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 654.757454] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e01696d-b4de-4af9-9abd-0621b08ca068 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.765294] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 654.768726] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a5b2b17-11e8-4255-bf18-680233b290bd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.776153] env[61728]: DEBUG oslo_vmware.api [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463817, 'name': PowerOnVM_Task, 'duration_secs': 0.997843} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.777662] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 654.777937] env[61728]: INFO nova.compute.manager [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Took 10.91 seconds to spawn the instance on the hypervisor. [ 654.778175] env[61728]: DEBUG nova.compute.manager [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 654.778543] env[61728]: DEBUG oslo_vmware.api [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Waiting for the task: (returnval){ [ 654.778543] env[61728]: value = "task-463822" [ 654.778543] env[61728]: _type = "Task" [ 654.778543] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.779603] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a45ff1d-ecdc-4ddd-92ea-0fe90a7d1845 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.797169] env[61728]: DEBUG oslo_vmware.api [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463822, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.187347] env[61728]: DEBUG oslo_vmware.api [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Task: {'id': task-463820, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.229527} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.187347] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 655.187550] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 655.187690] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 655.187872] env[61728]: INFO nova.compute.manager [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Took 1.11 seconds to destroy the instance on the hypervisor. [ 655.188151] env[61728]: DEBUG oslo.service.loopingcall [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.188339] env[61728]: DEBUG nova.compute.manager [-] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.188430] env[61728]: DEBUG nova.network.neutron [-] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 655.210361] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance b563e521-0e0e-45ea-bd5a-056285ea35b9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 655.253571] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463821, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.293068] env[61728]: DEBUG oslo_vmware.api [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463822, 'name': PowerOffVM_Task, 'duration_secs': 0.246829} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.293348] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 655.293529] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 655.293799] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d0f2c5e-68f1-4305-942f-1969ba463948 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.309628] env[61728]: INFO nova.compute.manager [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Took 49.33 seconds to build instance. [ 655.357298] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 655.357298] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 655.357298] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Deleting the datastore file [datastore1] 4fcc957e-5175-4589-b4da-d8451d92c815 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 655.357298] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ce5c8b1b-18cb-4f53-bd1d-5f6a4ba62312 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.366030] env[61728]: DEBUG oslo_vmware.api [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Waiting for the task: (returnval){ [ 655.366030] env[61728]: value = "task-463824" [ 655.366030] env[61728]: _type = "Task" [ 655.366030] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.372655] env[61728]: DEBUG oslo_vmware.api [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463824, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.714347] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 3bfefcfc-db97-4a9d-86cb-9fb1d8158863 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 655.753505] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463821, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.597893} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.753834] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 8c53c9c9-cdc1-407b-8827-8409cf137235/8c53c9c9-cdc1-407b-8827-8409cf137235.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 655.754105] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 655.754389] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0d6ffac-308a-4d39-ae20-a41eda316f1b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.761141] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Waiting for the task: (returnval){ [ 655.761141] env[61728]: value = "task-463825" [ 655.761141] env[61728]: _type = "Task" [ 655.761141] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.769685] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463825, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.812312] env[61728]: DEBUG oslo_concurrency.lockutils [None req-56ae5b4d-4f86-48f6-9cc2-2d7f90cfdde0 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Lock "69d7af16-7708-4df6-acca-093b6c1df1ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.830s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.874991] env[61728]: DEBUG oslo_vmware.api [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Task: {'id': task-463824, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.343957} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.875376] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 655.875575] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 655.875760] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 655.875934] env[61728]: INFO nova.compute.manager [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Took 1.12 seconds to destroy the instance on the hypervisor. [ 655.876209] env[61728]: DEBUG oslo.service.loopingcall [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.876402] env[61728]: DEBUG nova.compute.manager [-] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.876498] env[61728]: DEBUG nova.network.neutron [-] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 655.962255] env[61728]: DEBUG nova.network.neutron [-] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.217848] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance c14034b5-330b-4fb1-af31-d530e3231992 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.271281] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463825, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074033} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.271577] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 656.272497] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd8d829-5333-4e8f-8288-063d1cbae71f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.296108] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] 8c53c9c9-cdc1-407b-8827-8409cf137235/8c53c9c9-cdc1-407b-8827-8409cf137235.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 656.296742] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9a0ee16-889d-470e-beba-bbfded329b84 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.313857] env[61728]: DEBUG nova.compute.manager [req-5a41a233-96f7-48ab-b230-f95725c6c5be req-6dde8aca-81ef-4833-9409-bd77fe903793 service nova] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Received event network-vif-deleted-502b04b6-ffe9-4337-9574-426293d93c4c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 656.314077] env[61728]: DEBUG nova.compute.manager [req-5a41a233-96f7-48ab-b230-f95725c6c5be req-6dde8aca-81ef-4833-9409-bd77fe903793 service nova] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Received event network-vif-deleted-16958edb-78a0-4245-b286-406e92c6e241 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 656.314250] env[61728]: INFO nova.compute.manager [req-5a41a233-96f7-48ab-b230-f95725c6c5be req-6dde8aca-81ef-4833-9409-bd77fe903793 service nova] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Neutron deleted interface 16958edb-78a0-4245-b286-406e92c6e241; detaching it from the instance and deleting it from the info cache [ 656.314414] env[61728]: DEBUG nova.network.neutron [req-5a41a233-96f7-48ab-b230-f95725c6c5be req-6dde8aca-81ef-4833-9409-bd77fe903793 service nova] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.316151] env[61728]: DEBUG nova.compute.manager [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 656.323077] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Waiting for the task: (returnval){ [ 656.323077] env[61728]: value = "task-463826" [ 656.323077] env[61728]: _type = "Task" [ 656.323077] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.331031] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463826, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.464804] env[61728]: INFO nova.compute.manager [-] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Took 1.28 seconds to deallocate network for instance. [ 656.535275] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Acquiring lock "69d7af16-7708-4df6-acca-093b6c1df1ba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.535564] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Lock "69d7af16-7708-4df6-acca-093b6c1df1ba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.535812] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Acquiring lock "69d7af16-7708-4df6-acca-093b6c1df1ba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.536028] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Lock "69d7af16-7708-4df6-acca-093b6c1df1ba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.536220] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Lock "69d7af16-7708-4df6-acca-093b6c1df1ba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.538857] env[61728]: INFO nova.compute.manager [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Terminating instance [ 656.540280] env[61728]: DEBUG nova.compute.manager [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 656.540486] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 656.541334] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2331cde-d376-45cb-9444-9845c1c8de27 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.549296] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 656.549296] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b697dd16-4688-49ec-8a04-6cbb3af31daf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.556137] env[61728]: DEBUG oslo_vmware.api [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Waiting for the task: (returnval){ [ 656.556137] env[61728]: value = "task-463827" [ 656.556137] env[61728]: _type = "Task" [ 656.556137] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.565796] env[61728]: DEBUG oslo_vmware.api [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463827, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.614117] env[61728]: DEBUG nova.network.neutron [-] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.721955] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 223b1f93-d33e-4566-8c8e-75234a491814 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.822158] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-836c0df4-1f8c-4487-9124-e50bae2697ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.836335] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463826, 'name': ReconfigVM_Task, 'duration_secs': 0.248195} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.837787] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Reconfigured VM instance instance-00000025 to attach disk [datastore1] 8c53c9c9-cdc1-407b-8827-8409cf137235/8c53c9c9-cdc1-407b-8827-8409cf137235.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 656.838968] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a0ffe98-d288-447a-83f5-9a5473e7884b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.843728] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ac83e4-8dc3-475f-842a-99edf757ec8f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.855504] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.860478] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Waiting for the task: (returnval){ [ 656.860478] env[61728]: value = "task-463828" [ 656.860478] env[61728]: _type = "Task" [ 656.860478] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.879876] env[61728]: DEBUG nova.compute.manager [req-5a41a233-96f7-48ab-b230-f95725c6c5be req-6dde8aca-81ef-4833-9409-bd77fe903793 service nova] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Detach interface failed, port_id=16958edb-78a0-4245-b286-406e92c6e241, reason: Instance 4fcc957e-5175-4589-b4da-d8451d92c815 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 656.883904] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463828, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.974030] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.066027] env[61728]: DEBUG oslo_vmware.api [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463827, 'name': PowerOffVM_Task, 'duration_secs': 0.295794} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.066426] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 657.066611] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 657.066891] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25d59983-235b-403f-8c88-2092c5355931 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.117360] env[61728]: INFO nova.compute.manager [-] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Took 1.24 seconds to deallocate network for instance. [ 657.138043] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 657.138304] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 657.138494] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Deleting the datastore file [datastore1] 69d7af16-7708-4df6-acca-093b6c1df1ba {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 657.138780] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b8388a9-9897-45e7-ac87-b0e10289faf9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.146115] env[61728]: DEBUG oslo_vmware.api [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Waiting for the task: (returnval){ [ 657.146115] env[61728]: value = "task-463830" [ 657.146115] env[61728]: _type = "Task" [ 657.146115] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.154715] env[61728]: DEBUG oslo_vmware.api [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463830, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.224187] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.371353] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463828, 'name': Rename_Task, 'duration_secs': 0.149543} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.371608] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 657.371864] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cde2ae82-7fb8-47a9-87b5-0e18654e77d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.377727] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Waiting for the task: (returnval){ [ 657.377727] env[61728]: value = "task-463831" [ 657.377727] env[61728]: _type = "Task" [ 657.377727] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.385312] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463831, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.628388] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.655876] env[61728]: DEBUG oslo_vmware.api [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Task: {'id': task-463830, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.298582} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.656182] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 657.656387] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 657.656586] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 657.656788] env[61728]: INFO nova.compute.manager [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Took 1.12 seconds to destroy the instance on the hypervisor. [ 657.657075] env[61728]: DEBUG oslo.service.loopingcall [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 657.657294] env[61728]: DEBUG nova.compute.manager [-] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.657402] env[61728]: DEBUG nova.network.neutron [-] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 657.726808] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 6df7b619-8cc4-4dd9-8596-22dc83234a8e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.888788] env[61728]: DEBUG oslo_vmware.api [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463831, 'name': PowerOnVM_Task, 'duration_secs': 0.458068} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.889609] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 657.890022] env[61728]: INFO nova.compute.manager [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Took 8.08 seconds to spawn the instance on the hypervisor. [ 657.890233] env[61728]: DEBUG nova.compute.manager [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 657.891253] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a28c5d67-4086-4541-9319-c0f2c7dec381 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.229786] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance c84bfcd0-b145-4675-8b0a-5e8f94f65098 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 658.329713] env[61728]: DEBUG nova.compute.manager [req-0bac8290-5dcc-4aa7-8ec1-4bf807d6e224 req-5715ec0f-26e5-4b52-8ec3-7f5cd7242ebe service nova] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Received event network-vif-deleted-aae5c541-394b-4d2d-ac32-f2550bb74e28 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 658.329946] env[61728]: INFO nova.compute.manager [req-0bac8290-5dcc-4aa7-8ec1-4bf807d6e224 req-5715ec0f-26e5-4b52-8ec3-7f5cd7242ebe service nova] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Neutron deleted interface aae5c541-394b-4d2d-ac32-f2550bb74e28; detaching it from the instance and deleting it from the info cache [ 658.330144] env[61728]: DEBUG nova.network.neutron [req-0bac8290-5dcc-4aa7-8ec1-4bf807d6e224 req-5715ec0f-26e5-4b52-8ec3-7f5cd7242ebe service nova] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.410865] env[61728]: INFO nova.compute.manager [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Took 44.56 seconds to build instance. [ 658.418631] env[61728]: DEBUG nova.network.neutron [-] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.732700] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 658.833110] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e62b599-dd31-4692-962d-72b46870ddb3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.843814] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb508807-146b-4c88-8196-e630a9f472a8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.878807] env[61728]: DEBUG nova.compute.manager [req-0bac8290-5dcc-4aa7-8ec1-4bf807d6e224 req-5715ec0f-26e5-4b52-8ec3-7f5cd7242ebe service nova] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Detach interface failed, port_id=aae5c541-394b-4d2d-ac32-f2550bb74e28, reason: Instance 69d7af16-7708-4df6-acca-093b6c1df1ba could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 658.913810] env[61728]: DEBUG oslo_concurrency.lockutils [None req-199cb223-b5be-4166-bde4-fb4020bffe1c tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Lock "8c53c9c9-cdc1-407b-8827-8409cf137235" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.103s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.921202] env[61728]: INFO nova.compute.manager [-] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Took 1.26 seconds to deallocate network for instance. [ 659.175540] env[61728]: DEBUG nova.compute.manager [req-8fb2eda5-be39-41ff-8e0f-886132d485b2 req-9a1b58ec-6fba-4465-9aca-99f9f42bb355 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Received event network-changed-8ee9bdfc-796a-498b-845f-02218a86801a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 659.175782] env[61728]: DEBUG nova.compute.manager [req-8fb2eda5-be39-41ff-8e0f-886132d485b2 req-9a1b58ec-6fba-4465-9aca-99f9f42bb355 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Refreshing instance network info cache due to event network-changed-8ee9bdfc-796a-498b-845f-02218a86801a. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 659.176000] env[61728]: DEBUG oslo_concurrency.lockutils [req-8fb2eda5-be39-41ff-8e0f-886132d485b2 req-9a1b58ec-6fba-4465-9aca-99f9f42bb355 service nova] Acquiring lock "refresh_cache-8c53c9c9-cdc1-407b-8827-8409cf137235" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.176216] env[61728]: DEBUG oslo_concurrency.lockutils [req-8fb2eda5-be39-41ff-8e0f-886132d485b2 req-9a1b58ec-6fba-4465-9aca-99f9f42bb355 service nova] Acquired lock "refresh_cache-8c53c9c9-cdc1-407b-8827-8409cf137235" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.176291] env[61728]: DEBUG nova.network.neutron [req-8fb2eda5-be39-41ff-8e0f-886132d485b2 req-9a1b58ec-6fba-4465-9aca-99f9f42bb355 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Refreshing network info cache for port 8ee9bdfc-796a-498b-845f-02218a86801a {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 659.238635] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance f2692e22-6c14-4bb3-a080-607f0731105d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 659.239188] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Total usable vcpus: 48, total allocated vcpus: 19 {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 659.239590] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4224MB phys_disk=200GB used_disk=17GB total_vcpus=48 used_vcpus=19 pci_stats=[] {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 659.419276] env[61728]: DEBUG nova.compute.manager [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 659.428594] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.595759] env[61728]: DEBUG oslo_vmware.rw_handles [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52db8589-979d-33f8-2e74-23b708bfebf8/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 659.596811] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac22ba2-a8b3-4925-b8c8-f74f76ce1014 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.603475] env[61728]: DEBUG oslo_vmware.rw_handles [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52db8589-979d-33f8-2e74-23b708bfebf8/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 659.603657] env[61728]: ERROR oslo_vmware.rw_handles [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52db8589-979d-33f8-2e74-23b708bfebf8/disk-0.vmdk due to incomplete transfer. [ 659.605940] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e27c96cb-0cda-4002-b281-2d9d055a652f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.612795] env[61728]: DEBUG oslo_vmware.rw_handles [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52db8589-979d-33f8-2e74-23b708bfebf8/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 659.612911] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Uploaded image 4d578dcb-5948-4a97-a32f-c1442844b66f to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 659.614425] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 659.614684] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-43d68dd9-60a1-4365-94be-f6686881e653 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.619898] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 659.619898] env[61728]: value = "task-463832" [ 659.619898] env[61728]: _type = "Task" [ 659.619898] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.628862] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463832, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.744996] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cbc61f0-00d4-43b0-b3be-1f019fe6964e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.754795] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a225faf-dbe9-448e-87c8-baa1d17eed63 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.791137] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0378d1-5663-4ac7-af5a-26f1cdefefa5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.800944] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d7748d-1a07-4d38-862c-565478715271 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.813947] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.917961] env[61728]: DEBUG nova.network.neutron [req-8fb2eda5-be39-41ff-8e0f-886132d485b2 req-9a1b58ec-6fba-4465-9aca-99f9f42bb355 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Updated VIF entry in instance network info cache for port 8ee9bdfc-796a-498b-845f-02218a86801a. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 659.918357] env[61728]: DEBUG nova.network.neutron [req-8fb2eda5-be39-41ff-8e0f-886132d485b2 req-9a1b58ec-6fba-4465-9aca-99f9f42bb355 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Updating instance_info_cache with network_info: [{"id": "8ee9bdfc-796a-498b-845f-02218a86801a", "address": "fa:16:3e:ac:f0:22", "network": {"id": "22454043-3114-444a-aed9-fe24700844e1", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1570154546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.233", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "828ba999707e48f6bd9c11cfc74b1b2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ee9bdfc-79", "ovs_interfaceid": "8ee9bdfc-796a-498b-845f-02218a86801a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.940061] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.130659] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463832, 'name': Destroy_Task, 'duration_secs': 0.356732} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.130977] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Destroyed the VM [ 660.131234] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 660.131493] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cb4802f0-bcfc-4005-ad70-606b15b01a97 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.137665] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 660.137665] env[61728]: value = "task-463833" [ 660.137665] env[61728]: _type = "Task" [ 660.137665] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.145495] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463833, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.317796] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 660.421411] env[61728]: DEBUG oslo_concurrency.lockutils [req-8fb2eda5-be39-41ff-8e0f-886132d485b2 req-9a1b58ec-6fba-4465-9aca-99f9f42bb355 service nova] Releasing lock "refresh_cache-8c53c9c9-cdc1-407b-8827-8409cf137235" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.647848] env[61728]: DEBUG oslo_vmware.api [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463833, 'name': RemoveSnapshot_Task, 'duration_secs': 0.465566} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.647848] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 660.647933] env[61728]: INFO nova.compute.manager [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Took 17.56 seconds to snapshot the instance on the hypervisor. [ 660.822911] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 660.823147] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.720s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.823435] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.103s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.825049] env[61728]: INFO nova.compute.claims [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 661.188473] env[61728]: DEBUG nova.compute.manager [None req-2fb1aed3-5d5c-449b-9f0c-3c0f928f6236 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Found 1 images (rotation: 2) {{(pid=61728) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 661.608671] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "interface-620bb472-c36e-4c56-acdf-42e02614856b-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.608950] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-620bb472-c36e-4c56-acdf-42e02614856b-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.609234] env[61728]: DEBUG nova.objects.instance [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'flavor' on Instance uuid 620bb472-c36e-4c56-acdf-42e02614856b {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 662.107939] env[61728]: DEBUG nova.compute.manager [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 662.108874] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da8d1f9-3379-4fdc-a251-7101ebf27d2d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.200572] env[61728]: DEBUG nova.objects.instance [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'pci_requests' on Instance uuid 620bb472-c36e-4c56-acdf-42e02614856b {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 662.269989] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9cebb06-f7cb-46eb-9431-1b4fd287931f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.277343] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c858439-f912-4b9e-bde6-1cdc683021f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.308482] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d931fe-e1a5-4c8b-acdb-adc1048a3677 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.315437] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-149b8d5f-a54d-4c84-80df-2e2999c16d4f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.328282] env[61728]: DEBUG nova.compute.provider_tree [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.625084] env[61728]: INFO nova.compute.manager [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] instance snapshotting [ 662.625830] env[61728]: DEBUG nova.objects.instance [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'flavor' on Instance uuid 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 662.705374] env[61728]: DEBUG nova.objects.base [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Object Instance<620bb472-c36e-4c56-acdf-42e02614856b> lazy-loaded attributes: flavor,pci_requests {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 662.705635] env[61728]: DEBUG nova.network.neutron [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 662.742517] env[61728]: DEBUG nova.policy [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e51efcf23db4d5496edbc3b7104b4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32bb35e1dfce40e48be08bb568d3f2b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 662.831953] env[61728]: DEBUG nova.scheduler.client.report [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.012283] env[61728]: DEBUG nova.network.neutron [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Successfully created port: f605e2a3-ef9e-4642-8d18-e43436149349 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 663.131612] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153fbd39-d160-4d62-ba3f-e5ccc4da07b2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.151515] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd3add66-256e-44ed-a727-ae8480788e59 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.337404] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.337954] env[61728]: DEBUG nova.compute.manager [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 663.340440] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.339s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.341821] env[61728]: INFO nova.compute.claims [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 663.661285] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 663.662016] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-174b3d43-b524-4811-b570-919920f2cb65 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.670682] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 663.670682] env[61728]: value = "task-463834" [ 663.670682] env[61728]: _type = "Task" [ 663.670682] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.678261] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463834, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.846502] env[61728]: DEBUG nova.compute.utils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 663.850029] env[61728]: DEBUG nova.compute.manager [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 663.850029] env[61728]: DEBUG nova.network.neutron [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 663.904689] env[61728]: DEBUG nova.policy [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f8738985bd14fc3a811876fa6041674', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '037e1941b7964a2b94d5b0f5c280956e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 664.172593] env[61728]: DEBUG nova.network.neutron [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Successfully created port: b5b01c97-ed6d-41f8-98e3-9489d2e04f12 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 664.182879] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463834, 'name': CreateSnapshot_Task, 'duration_secs': 0.428138} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.183164] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 664.183896] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7aa6d3a-fcda-4d3d-9b74-d3914981bd0b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.352951] env[61728]: DEBUG nova.compute.manager [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 664.568186] env[61728]: DEBUG nova.network.neutron [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Successfully created port: 48764907-e12d-424f-8b5b-20c70e244c6a {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 664.624753] env[61728]: DEBUG nova.compute.manager [req-d0c36bb8-4c1a-4ac1-9789-c0535b525d08 req-3bea7f5e-6560-4910-9613-2308720ef86e service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Received event network-vif-plugged-f605e2a3-ef9e-4642-8d18-e43436149349 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 664.625324] env[61728]: DEBUG oslo_concurrency.lockutils [req-d0c36bb8-4c1a-4ac1-9789-c0535b525d08 req-3bea7f5e-6560-4910-9613-2308720ef86e service nova] Acquiring lock "620bb472-c36e-4c56-acdf-42e02614856b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.626019] env[61728]: DEBUG oslo_concurrency.lockutils [req-d0c36bb8-4c1a-4ac1-9789-c0535b525d08 req-3bea7f5e-6560-4910-9613-2308720ef86e service nova] Lock "620bb472-c36e-4c56-acdf-42e02614856b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.626019] env[61728]: DEBUG oslo_concurrency.lockutils [req-d0c36bb8-4c1a-4ac1-9789-c0535b525d08 req-3bea7f5e-6560-4910-9613-2308720ef86e service nova] Lock "620bb472-c36e-4c56-acdf-42e02614856b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.626266] env[61728]: DEBUG nova.compute.manager [req-d0c36bb8-4c1a-4ac1-9789-c0535b525d08 req-3bea7f5e-6560-4910-9613-2308720ef86e service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] No waiting events found dispatching network-vif-plugged-f605e2a3-ef9e-4642-8d18-e43436149349 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 664.626593] env[61728]: WARNING nova.compute.manager [req-d0c36bb8-4c1a-4ac1-9789-c0535b525d08 req-3bea7f5e-6560-4910-9613-2308720ef86e service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Received unexpected event network-vif-plugged-f605e2a3-ef9e-4642-8d18-e43436149349 for instance with vm_state active and task_state None. [ 664.702461] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 664.705327] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-bb2fac25-667a-4058-9546-f414cdf56355 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.716424] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 664.716424] env[61728]: value = "task-463835" [ 664.716424] env[61728]: _type = "Task" [ 664.716424] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.727651] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463835, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.838050] env[61728]: DEBUG nova.network.neutron [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Successfully updated port: f605e2a3-ef9e-4642-8d18-e43436149349 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 664.936612] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a0eecf1-dcfc-405a-a5c4-a8c4030ecd9c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.944512] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62dc726c-b087-48e6-b007-46b3609cb5bb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.979889] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0076da61-a141-4454-a6cc-8c0afd1aba9a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.987494] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4bea9fc-a0d2-4985-adca-0f71292ad6cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.001685] env[61728]: DEBUG nova.compute.provider_tree [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.061660] env[61728]: DEBUG nova.network.neutron [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Successfully created port: b24101a4-7e85-4510-b0ce-1ebede43eb5c {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 665.228169] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463835, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.342501] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.342838] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.343113] env[61728]: DEBUG nova.network.neutron [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 665.361305] env[61728]: DEBUG nova.compute.manager [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 665.386366] env[61728]: DEBUG nova.virt.hardware [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 665.386678] env[61728]: DEBUG nova.virt.hardware [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 665.386877] env[61728]: DEBUG nova.virt.hardware [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 665.387120] env[61728]: DEBUG nova.virt.hardware [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 665.387321] env[61728]: DEBUG nova.virt.hardware [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 665.387537] env[61728]: DEBUG nova.virt.hardware [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 665.387789] env[61728]: DEBUG nova.virt.hardware [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 665.387989] env[61728]: DEBUG nova.virt.hardware [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 665.388226] env[61728]: DEBUG nova.virt.hardware [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 665.388469] env[61728]: DEBUG nova.virt.hardware [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 665.388713] env[61728]: DEBUG nova.virt.hardware [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 665.389953] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc50ffa-7b71-4435-8845-aff172c8560d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.397856] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aeece26-00fb-44ad-ae14-387a1d855b4e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.505473] env[61728]: DEBUG nova.scheduler.client.report [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 665.727708] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463835, 'name': CloneVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.887098] env[61728]: WARNING nova.network.neutron [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] 20ba9b61-3be9-4ad2-a1d0-88c810873bcb already exists in list: networks containing: ['20ba9b61-3be9-4ad2-a1d0-88c810873bcb']. ignoring it [ 666.011527] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.671s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.011917] env[61728]: DEBUG nova.compute.manager [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 666.014552] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.028s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.014710] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.016918] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.828s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.018310] env[61728]: INFO nova.compute.claims [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 666.048274] env[61728]: INFO nova.scheduler.client.report [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleted allocations for instance fc3c1d93-b095-4d80-8d71-243b66d85a46 [ 666.180766] env[61728]: DEBUG nova.network.neutron [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Updating instance_info_cache with network_info: [{"id": "edc90742-f02d-42ad-85cd-b4242f797c33", "address": "fa:16:3e:aa:75:a5", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedc90742-f0", "ovs_interfaceid": "edc90742-f02d-42ad-85cd-b4242f797c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f605e2a3-ef9e-4642-8d18-e43436149349", "address": "fa:16:3e:b7:de:93", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf605e2a3-ef", "ovs_interfaceid": "f605e2a3-ef9e-4642-8d18-e43436149349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.229946] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463835, 'name': CloneVM_Task, 'duration_secs': 1.019897} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.230238] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Created linked-clone VM from snapshot [ 666.231037] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0ebbfd-3c3b-4a86-843a-bdd9c1ce8d6a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.237955] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Uploading image b7db4964-4a64-4b56-bcc4-482c37cec071 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 666.261592] env[61728]: DEBUG oslo_vmware.rw_handles [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 666.261592] env[61728]: value = "vm-122034" [ 666.261592] env[61728]: _type = "VirtualMachine" [ 666.261592] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 666.262448] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-aea84bff-8aa3-49e7-bcdd-9c9c8290c760 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.268695] env[61728]: DEBUG oslo_vmware.rw_handles [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lease: (returnval){ [ 666.268695] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52efe7dd-6850-0d1f-03c0-244d2c5be92a" [ 666.268695] env[61728]: _type = "HttpNfcLease" [ 666.268695] env[61728]: } obtained for exporting VM: (result){ [ 666.268695] env[61728]: value = "vm-122034" [ 666.268695] env[61728]: _type = "VirtualMachine" [ 666.268695] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 666.268948] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the lease: (returnval){ [ 666.268948] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52efe7dd-6850-0d1f-03c0-244d2c5be92a" [ 666.268948] env[61728]: _type = "HttpNfcLease" [ 666.268948] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 666.275079] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 666.275079] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52efe7dd-6850-0d1f-03c0-244d2c5be92a" [ 666.275079] env[61728]: _type = "HttpNfcLease" [ 666.275079] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 666.499078] env[61728]: DEBUG nova.network.neutron [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Successfully updated port: b5b01c97-ed6d-41f8-98e3-9489d2e04f12 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 666.523177] env[61728]: DEBUG nova.compute.utils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 666.528183] env[61728]: DEBUG nova.compute.manager [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 666.528490] env[61728]: DEBUG nova.network.neutron [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 666.558985] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e0d8ee6-141c-4286-8798-13a2ca7bf6b0 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "fc3c1d93-b095-4d80-8d71-243b66d85a46" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.152s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.565748] env[61728]: DEBUG nova.policy [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd32b5118a94042f797dea1fdf6bd05b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08367aaceba548fe93faaedf6371817d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 666.684526] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.685219] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.685384] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.686562] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc6ce3c-6db8-4dcf-8566-1b5a523972d4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.706885] env[61728]: DEBUG nova.virt.hardware [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 666.707172] env[61728]: DEBUG nova.virt.hardware [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 666.707327] env[61728]: DEBUG nova.virt.hardware [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 666.707525] env[61728]: DEBUG nova.virt.hardware [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 666.707681] env[61728]: DEBUG nova.virt.hardware [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 666.707849] env[61728]: DEBUG nova.virt.hardware [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 666.708180] env[61728]: DEBUG nova.virt.hardware [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 666.708359] env[61728]: DEBUG nova.virt.hardware [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 666.708577] env[61728]: DEBUG nova.virt.hardware [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 666.708781] env[61728]: DEBUG nova.virt.hardware [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 666.708976] env[61728]: DEBUG nova.virt.hardware [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 666.716665] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Reconfiguring VM to attach interface {{(pid=61728) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 666.717010] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0fe371e1-73b5-45f4-b689-7ea2bac1bab2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.737402] env[61728]: DEBUG oslo_vmware.api [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 666.737402] env[61728]: value = "task-463837" [ 666.737402] env[61728]: _type = "Task" [ 666.737402] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.744203] env[61728]: DEBUG oslo_vmware.api [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463837, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.780747] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 666.780747] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52efe7dd-6850-0d1f-03c0-244d2c5be92a" [ 666.780747] env[61728]: _type = "HttpNfcLease" [ 666.780747] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 666.781089] env[61728]: DEBUG oslo_vmware.rw_handles [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 666.781089] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52efe7dd-6850-0d1f-03c0-244d2c5be92a" [ 666.781089] env[61728]: _type = "HttpNfcLease" [ 666.781089] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 666.781886] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22101a14-6d65-4457-97ea-9409551fc9b3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.789573] env[61728]: DEBUG oslo_vmware.rw_handles [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52cf9ca3-103c-0d0d-1b2a-df57638b0204/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 666.789795] env[61728]: DEBUG oslo_vmware.rw_handles [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52cf9ca3-103c-0d0d-1b2a-df57638b0204/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 666.849608] env[61728]: DEBUG nova.network.neutron [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Successfully created port: c170e82d-b764-4616-863c-35d45b72edbc {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 666.869668] env[61728]: DEBUG nova.compute.manager [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Received event network-changed-f605e2a3-ef9e-4642-8d18-e43436149349 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 666.870017] env[61728]: DEBUG nova.compute.manager [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Refreshing instance network info cache due to event network-changed-f605e2a3-ef9e-4642-8d18-e43436149349. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 666.870395] env[61728]: DEBUG oslo_concurrency.lockutils [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] Acquiring lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.870611] env[61728]: DEBUG oslo_concurrency.lockutils [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] Acquired lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.870840] env[61728]: DEBUG nova.network.neutron [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Refreshing network info cache for port f605e2a3-ef9e-4642-8d18-e43436149349 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 666.882597] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-035486cf-6da9-4314-bffc-0fcaa492a7f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.028061] env[61728]: DEBUG nova.compute.manager [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 667.251804] env[61728]: DEBUG oslo_vmware.api [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.628514] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c32d714-eafd-4bb3-b6f9-64ee9c853b17 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.632197] env[61728]: DEBUG nova.network.neutron [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Updated VIF entry in instance network info cache for port f605e2a3-ef9e-4642-8d18-e43436149349. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 667.632663] env[61728]: DEBUG nova.network.neutron [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Updating instance_info_cache with network_info: [{"id": "edc90742-f02d-42ad-85cd-b4242f797c33", "address": "fa:16:3e:aa:75:a5", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedc90742-f0", "ovs_interfaceid": "edc90742-f02d-42ad-85cd-b4242f797c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f605e2a3-ef9e-4642-8d18-e43436149349", "address": "fa:16:3e:b7:de:93", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf605e2a3-ef", "ovs_interfaceid": "f605e2a3-ef9e-4642-8d18-e43436149349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.639000] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852a705b-10c6-4685-ae7a-03e55625d0e7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.671646] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26dc6bc7-dcb3-472e-8917-0ebf37114ac7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.679369] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086ade3a-ff97-4e90-8fe3-0b9ca636e08c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.694353] env[61728]: DEBUG nova.compute.provider_tree [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.747663] env[61728]: DEBUG oslo_vmware.api [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463837, 'name': ReconfigVM_Task, 'duration_secs': 0.629572} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.748222] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.748444] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Reconfigured VM to attach interface {{(pid=61728) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 668.044589] env[61728]: DEBUG nova.compute.manager [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 668.067955] env[61728]: DEBUG nova.virt.hardware [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 668.068259] env[61728]: DEBUG nova.virt.hardware [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 668.068528] env[61728]: DEBUG nova.virt.hardware [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 668.068721] env[61728]: DEBUG nova.virt.hardware [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 668.068874] env[61728]: DEBUG nova.virt.hardware [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 668.069041] env[61728]: DEBUG nova.virt.hardware [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 668.069260] env[61728]: DEBUG nova.virt.hardware [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 668.069425] env[61728]: DEBUG nova.virt.hardware [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 668.069610] env[61728]: DEBUG nova.virt.hardware [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 668.069779] env[61728]: DEBUG nova.virt.hardware [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 668.069955] env[61728]: DEBUG nova.virt.hardware [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.070835] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e69f4930-4bf9-41da-830f-65381f379bef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.078772] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e326bf-730c-4f6d-a98a-fbe7bb283aea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.136288] env[61728]: DEBUG oslo_concurrency.lockutils [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] Releasing lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.136869] env[61728]: DEBUG nova.compute.manager [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Received event network-vif-plugged-b5b01c97-ed6d-41f8-98e3-9489d2e04f12 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 668.137256] env[61728]: DEBUG oslo_concurrency.lockutils [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] Acquiring lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.137571] env[61728]: DEBUG oslo_concurrency.lockutils [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] Lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.137748] env[61728]: DEBUG oslo_concurrency.lockutils [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] Lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.137921] env[61728]: DEBUG nova.compute.manager [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] No waiting events found dispatching network-vif-plugged-b5b01c97-ed6d-41f8-98e3-9489d2e04f12 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 668.138107] env[61728]: WARNING nova.compute.manager [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Received unexpected event network-vif-plugged-b5b01c97-ed6d-41f8-98e3-9489d2e04f12 for instance with vm_state building and task_state spawning. [ 668.138292] env[61728]: DEBUG nova.compute.manager [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Received event network-changed-b5b01c97-ed6d-41f8-98e3-9489d2e04f12 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 668.138454] env[61728]: DEBUG nova.compute.manager [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Refreshing instance network info cache due to event network-changed-b5b01c97-ed6d-41f8-98e3-9489d2e04f12. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 668.138721] env[61728]: DEBUG oslo_concurrency.lockutils [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] Acquiring lock "refresh_cache-3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.138873] env[61728]: DEBUG oslo_concurrency.lockutils [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] Acquired lock "refresh_cache-3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.139060] env[61728]: DEBUG nova.network.neutron [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Refreshing network info cache for port b5b01c97-ed6d-41f8-98e3-9489d2e04f12 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 668.200123] env[61728]: DEBUG nova.scheduler.client.report [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.254048] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8307765-bdb3-4982-9523-eddc322fe8b2 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-620bb472-c36e-4c56-acdf-42e02614856b-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.645s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.543159] env[61728]: DEBUG nova.network.neutron [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Successfully updated port: c170e82d-b764-4616-863c-35d45b72edbc {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 668.654534] env[61728]: DEBUG nova.network.neutron [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Successfully updated port: 48764907-e12d-424f-8b5b-20c70e244c6a {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 668.681374] env[61728]: DEBUG nova.network.neutron [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.704456] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.687s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.704996] env[61728]: DEBUG nova.compute.manager [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 668.707599] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.530s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.709187] env[61728]: INFO nova.compute.claims [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 668.780440] env[61728]: DEBUG nova.network.neutron [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.046694] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "refresh_cache-507b5aba-aec9-438b-92c0-5efa17959bc7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.046880] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "refresh_cache-507b5aba-aec9-438b-92c0-5efa17959bc7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.047036] env[61728]: DEBUG nova.network.neutron [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 669.182703] env[61728]: DEBUG nova.compute.manager [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Received event network-vif-plugged-c170e82d-b764-4616-863c-35d45b72edbc {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 669.182955] env[61728]: DEBUG oslo_concurrency.lockutils [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] Acquiring lock "507b5aba-aec9-438b-92c0-5efa17959bc7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.183228] env[61728]: DEBUG oslo_concurrency.lockutils [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] Lock "507b5aba-aec9-438b-92c0-5efa17959bc7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.183427] env[61728]: DEBUG oslo_concurrency.lockutils [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] Lock "507b5aba-aec9-438b-92c0-5efa17959bc7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.183619] env[61728]: DEBUG nova.compute.manager [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] No waiting events found dispatching network-vif-plugged-c170e82d-b764-4616-863c-35d45b72edbc {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 669.183805] env[61728]: WARNING nova.compute.manager [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Received unexpected event network-vif-plugged-c170e82d-b764-4616-863c-35d45b72edbc for instance with vm_state building and task_state spawning. [ 669.184064] env[61728]: DEBUG nova.compute.manager [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Received event network-changed-c170e82d-b764-4616-863c-35d45b72edbc {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 669.184283] env[61728]: DEBUG nova.compute.manager [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Refreshing instance network info cache due to event network-changed-c170e82d-b764-4616-863c-35d45b72edbc. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 669.184463] env[61728]: DEBUG oslo_concurrency.lockutils [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] Acquiring lock "refresh_cache-507b5aba-aec9-438b-92c0-5efa17959bc7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.216048] env[61728]: DEBUG nova.compute.utils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 669.221022] env[61728]: DEBUG nova.compute.manager [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 669.221022] env[61728]: DEBUG nova.network.neutron [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 669.286876] env[61728]: DEBUG oslo_concurrency.lockutils [req-345204e4-32e4-4fa8-9c3d-1c2b80e2fae9 req-3d5db4d6-7273-472b-9ac1-3814d4822a53 service nova] Releasing lock "refresh_cache-3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.304802] env[61728]: DEBUG nova.policy [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2a4fae9faff24030b6ffead1ec9b252a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f6788ff3c6ac4aacad4af5e1ba42fb03', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 669.602602] env[61728]: DEBUG nova.network.neutron [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.722778] env[61728]: DEBUG nova.compute.manager [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 669.787486] env[61728]: DEBUG nova.network.neutron [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Successfully created port: fdd5e1d2-60d3-4304-a31b-e6b680c01c63 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 669.863235] env[61728]: DEBUG nova.network.neutron [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Updating instance_info_cache with network_info: [{"id": "c170e82d-b764-4616-863c-35d45b72edbc", "address": "fa:16:3e:f1:7b:cd", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc170e82d-b7", "ovs_interfaceid": "c170e82d-b764-4616-863c-35d45b72edbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.261043] env[61728]: DEBUG oslo_concurrency.lockutils [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "interface-620bb472-c36e-4c56-acdf-42e02614856b-f605e2a3-ef9e-4642-8d18-e43436149349" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.261043] env[61728]: DEBUG oslo_concurrency.lockutils [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-620bb472-c36e-4c56-acdf-42e02614856b-f605e2a3-ef9e-4642-8d18-e43436149349" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.294637] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9068b9e4-ebdf-4a77-88d2-94b26f19284e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.303207] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2382f90c-3ac3-49f8-abbb-43705bb57ace {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.335756] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76906e1d-560b-4cf9-a800-630331dc6998 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.343375] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c4444fb-398a-4fa4-bc9a-85c1065b4817 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.356504] env[61728]: DEBUG nova.compute.provider_tree [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.365647] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "refresh_cache-507b5aba-aec9-438b-92c0-5efa17959bc7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.366029] env[61728]: DEBUG nova.compute.manager [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Instance network_info: |[{"id": "c170e82d-b764-4616-863c-35d45b72edbc", "address": "fa:16:3e:f1:7b:cd", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc170e82d-b7", "ovs_interfaceid": "c170e82d-b764-4616-863c-35d45b72edbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 670.366940] env[61728]: DEBUG oslo_concurrency.lockutils [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] Acquired lock "refresh_cache-507b5aba-aec9-438b-92c0-5efa17959bc7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.367148] env[61728]: DEBUG nova.network.neutron [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Refreshing network info cache for port c170e82d-b764-4616-863c-35d45b72edbc {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 670.368277] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:7b:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd733acc2-07d0-479e-918c-ec8a21925389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c170e82d-b764-4616-863c-35d45b72edbc', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 670.376062] env[61728]: DEBUG oslo.service.loopingcall [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 670.377053] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 670.377300] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-297a1d2a-6feb-4a61-8879-7b33acf2eea5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.397244] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 670.397244] env[61728]: value = "task-463838" [ 670.397244] env[61728]: _type = "Task" [ 670.397244] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.405433] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463838, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.734742] env[61728]: DEBUG nova.compute.manager [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 670.762439] env[61728]: DEBUG nova.virt.hardware [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 670.763271] env[61728]: DEBUG nova.virt.hardware [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 670.763271] env[61728]: DEBUG nova.virt.hardware [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 670.763271] env[61728]: DEBUG nova.virt.hardware [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 670.763489] env[61728]: DEBUG nova.virt.hardware [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 670.763536] env[61728]: DEBUG nova.virt.hardware [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 670.763755] env[61728]: DEBUG nova.virt.hardware [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 670.763929] env[61728]: DEBUG nova.virt.hardware [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 670.764306] env[61728]: DEBUG nova.virt.hardware [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 670.764800] env[61728]: DEBUG nova.virt.hardware [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 670.765134] env[61728]: DEBUG nova.virt.hardware [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 670.765630] env[61728]: DEBUG oslo_concurrency.lockutils [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.765793] env[61728]: DEBUG oslo_concurrency.lockutils [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.767027] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae790917-301e-4662-803b-b3b76b09ae52 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.772786] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af21ee6f-a9a2-4f21-a8b2-8026408b02a0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.791679] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e95e54-c493-4b27-ba83-0b6a6c0e897b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.795447] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9acb6389-8e79-4f50-a3f0-c6d686b8141a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.832538] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Reconfiguring VM to detach interface {{(pid=61728) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 670.833483] env[61728]: DEBUG nova.network.neutron [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Successfully updated port: b24101a4-7e85-4510-b0ce-1ebede43eb5c {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 670.834892] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16f0e139-5129-481b-9580-f1611b7e8682 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.856080] env[61728]: DEBUG oslo_vmware.api [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 670.856080] env[61728]: value = "task-463839" [ 670.856080] env[61728]: _type = "Task" [ 670.856080] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.862897] env[61728]: DEBUG nova.scheduler.client.report [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.872226] env[61728]: DEBUG oslo_vmware.api [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.907328] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463838, 'name': CreateVM_Task, 'duration_secs': 0.369939} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.909494] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 670.910342] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.910434] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.910681] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 670.911252] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ca79b74-3671-45e4-9fda-1b530c880ce2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.916238] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 670.916238] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5274fd0c-0568-1f86-69d5-49e11390becd" [ 670.916238] env[61728]: _type = "Task" [ 670.916238] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.924354] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5274fd0c-0568-1f86-69d5-49e11390becd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.089790] env[61728]: DEBUG nova.network.neutron [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Updated VIF entry in instance network info cache for port c170e82d-b764-4616-863c-35d45b72edbc. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 671.090381] env[61728]: DEBUG nova.network.neutron [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Updating instance_info_cache with network_info: [{"id": "c170e82d-b764-4616-863c-35d45b72edbc", "address": "fa:16:3e:f1:7b:cd", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc170e82d-b7", "ovs_interfaceid": "c170e82d-b764-4616-863c-35d45b72edbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.208772] env[61728]: DEBUG nova.compute.manager [req-fc98597a-9310-4419-b3dc-16d259915ffa req-3b8d9699-c9d5-4a3d-a03a-ef7563c76559 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Received event network-vif-plugged-b24101a4-7e85-4510-b0ce-1ebede43eb5c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 671.209198] env[61728]: DEBUG oslo_concurrency.lockutils [req-fc98597a-9310-4419-b3dc-16d259915ffa req-3b8d9699-c9d5-4a3d-a03a-ef7563c76559 service nova] Acquiring lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.209482] env[61728]: DEBUG oslo_concurrency.lockutils [req-fc98597a-9310-4419-b3dc-16d259915ffa req-3b8d9699-c9d5-4a3d-a03a-ef7563c76559 service nova] Lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.209741] env[61728]: DEBUG oslo_concurrency.lockutils [req-fc98597a-9310-4419-b3dc-16d259915ffa req-3b8d9699-c9d5-4a3d-a03a-ef7563c76559 service nova] Lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.209974] env[61728]: DEBUG nova.compute.manager [req-fc98597a-9310-4419-b3dc-16d259915ffa req-3b8d9699-c9d5-4a3d-a03a-ef7563c76559 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] No waiting events found dispatching network-vif-plugged-b24101a4-7e85-4510-b0ce-1ebede43eb5c {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 671.210206] env[61728]: WARNING nova.compute.manager [req-fc98597a-9310-4419-b3dc-16d259915ffa req-3b8d9699-c9d5-4a3d-a03a-ef7563c76559 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Received unexpected event network-vif-plugged-b24101a4-7e85-4510-b0ce-1ebede43eb5c for instance with vm_state building and task_state spawning. [ 671.210438] env[61728]: DEBUG nova.compute.manager [req-fc98597a-9310-4419-b3dc-16d259915ffa req-3b8d9699-c9d5-4a3d-a03a-ef7563c76559 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Received event network-changed-b24101a4-7e85-4510-b0ce-1ebede43eb5c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 671.210666] env[61728]: DEBUG nova.compute.manager [req-fc98597a-9310-4419-b3dc-16d259915ffa req-3b8d9699-c9d5-4a3d-a03a-ef7563c76559 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Refreshing instance network info cache due to event network-changed-b24101a4-7e85-4510-b0ce-1ebede43eb5c. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 671.210918] env[61728]: DEBUG oslo_concurrency.lockutils [req-fc98597a-9310-4419-b3dc-16d259915ffa req-3b8d9699-c9d5-4a3d-a03a-ef7563c76559 service nova] Acquiring lock "refresh_cache-3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.211142] env[61728]: DEBUG oslo_concurrency.lockutils [req-fc98597a-9310-4419-b3dc-16d259915ffa req-3b8d9699-c9d5-4a3d-a03a-ef7563c76559 service nova] Acquired lock "refresh_cache-3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.211358] env[61728]: DEBUG nova.network.neutron [req-fc98597a-9310-4419-b3dc-16d259915ffa req-3b8d9699-c9d5-4a3d-a03a-ef7563c76559 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Refreshing network info cache for port b24101a4-7e85-4510-b0ce-1ebede43eb5c {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 671.328425] env[61728]: DEBUG nova.network.neutron [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Successfully updated port: fdd5e1d2-60d3-4304-a31b-e6b680c01c63 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 671.350310] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "refresh_cache-3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.368829] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.661s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.369382] env[61728]: DEBUG nova.compute.manager [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 671.371827] env[61728]: DEBUG oslo_vmware.api [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.372312] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.427s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.373778] env[61728]: INFO nova.compute.claims [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 671.430146] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5274fd0c-0568-1f86-69d5-49e11390becd, 'name': SearchDatastore_Task, 'duration_secs': 0.012758} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.430146] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.430146] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 671.430146] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.430146] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.430146] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 671.430146] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23eafc4f-bdae-44b1-ab22-117f4ddb8428 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.438051] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 671.438051] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 671.438508] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-671a7cbf-0b91-42cc-bdc9-0354e0b25265 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.444125] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 671.444125] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52581a11-a8f5-b1bc-27e4-f0335a389f86" [ 671.444125] env[61728]: _type = "Task" [ 671.444125] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.451920] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52581a11-a8f5-b1bc-27e4-f0335a389f86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.593445] env[61728]: DEBUG oslo_concurrency.lockutils [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] Releasing lock "refresh_cache-507b5aba-aec9-438b-92c0-5efa17959bc7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.593759] env[61728]: DEBUG nova.compute.manager [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Received event network-vif-plugged-48764907-e12d-424f-8b5b-20c70e244c6a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 671.593983] env[61728]: DEBUG oslo_concurrency.lockutils [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] Acquiring lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.594224] env[61728]: DEBUG oslo_concurrency.lockutils [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] Lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.594764] env[61728]: DEBUG oslo_concurrency.lockutils [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] Lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.594764] env[61728]: DEBUG nova.compute.manager [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] No waiting events found dispatching network-vif-plugged-48764907-e12d-424f-8b5b-20c70e244c6a {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 671.594764] env[61728]: WARNING nova.compute.manager [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Received unexpected event network-vif-plugged-48764907-e12d-424f-8b5b-20c70e244c6a for instance with vm_state building and task_state spawning. [ 671.594903] env[61728]: DEBUG nova.compute.manager [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Received event network-changed-48764907-e12d-424f-8b5b-20c70e244c6a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 671.595058] env[61728]: DEBUG nova.compute.manager [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Refreshing instance network info cache due to event network-changed-48764907-e12d-424f-8b5b-20c70e244c6a. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 671.595240] env[61728]: DEBUG oslo_concurrency.lockutils [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] Acquiring lock "refresh_cache-3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.747252] env[61728]: DEBUG nova.network.neutron [req-fc98597a-9310-4419-b3dc-16d259915ffa req-3b8d9699-c9d5-4a3d-a03a-ef7563c76559 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.831365] env[61728]: DEBUG nova.network.neutron [req-fc98597a-9310-4419-b3dc-16d259915ffa req-3b8d9699-c9d5-4a3d-a03a-ef7563c76559 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.832940] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "refresh_cache-650d2875-68f7-4131-b9b7-9799ee08f1d6" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.833097] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquired lock "refresh_cache-650d2875-68f7-4131-b9b7-9799ee08f1d6" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.833241] env[61728]: DEBUG nova.network.neutron [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 671.867194] env[61728]: DEBUG oslo_vmware.api [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.873770] env[61728]: DEBUG nova.compute.utils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 671.875278] env[61728]: DEBUG nova.compute.manager [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 671.875451] env[61728]: DEBUG nova.network.neutron [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 671.920114] env[61728]: DEBUG nova.policy [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e2344b0afa047f89e7623b8cc7a7fdc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad57e3e090854cee813c4dfd8eea26ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 671.955978] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52581a11-a8f5-b1bc-27e4-f0335a389f86, 'name': SearchDatastore_Task, 'duration_secs': 0.011145} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.957885] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9eb37c6-2c83-441c-9a79-c4177d7b5e99 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.965592] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 671.965592] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524db04d-d7ec-59e8-2f0b-98344e0e9a20" [ 671.965592] env[61728]: _type = "Task" [ 671.965592] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.974333] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524db04d-d7ec-59e8-2f0b-98344e0e9a20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.189171] env[61728]: DEBUG nova.network.neutron [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Successfully created port: 2537e303-47b7-4ed7-b8b9-7fd10ffc69d5 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 672.338903] env[61728]: DEBUG oslo_concurrency.lockutils [req-fc98597a-9310-4419-b3dc-16d259915ffa req-3b8d9699-c9d5-4a3d-a03a-ef7563c76559 service nova] Releasing lock "refresh_cache-3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.339333] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquired lock "refresh_cache-3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.339588] env[61728]: DEBUG nova.network.neutron [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 672.368475] env[61728]: DEBUG oslo_vmware.api [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.379247] env[61728]: DEBUG nova.network.neutron [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.381406] env[61728]: DEBUG nova.compute.manager [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 672.486011] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524db04d-d7ec-59e8-2f0b-98344e0e9a20, 'name': SearchDatastore_Task, 'duration_secs': 0.013186} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.486343] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.486731] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 507b5aba-aec9-438b-92c0-5efa17959bc7/507b5aba-aec9-438b-92c0-5efa17959bc7.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 672.486903] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-91a21811-00cb-4f82-8e6c-d2e191d71826 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.497081] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 672.497081] env[61728]: value = "task-463840" [ 672.497081] env[61728]: _type = "Task" [ 672.497081] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.508703] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463840, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.561161] env[61728]: DEBUG nova.network.neutron [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Updating instance_info_cache with network_info: [{"id": "fdd5e1d2-60d3-4304-a31b-e6b680c01c63", "address": "fa:16:3e:1b:8d:97", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.126", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdd5e1d2-60", "ovs_interfaceid": "fdd5e1d2-60d3-4304-a31b-e6b680c01c63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.873347] env[61728]: DEBUG oslo_vmware.api [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.897477] env[61728]: DEBUG nova.network.neutron [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.943133] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981303da-f6fd-48aa-97c7-0a3ef2dc3dad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.956273] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ab40de-3f5a-4f2a-9114-edb15dec0bc5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.995846] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8846acfc-9317-4cff-a1da-9711b4c0d4d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.009734] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463840, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.013260] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ed5ed3-9906-46ec-b36d-30d229c0bb3e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.032386] env[61728]: DEBUG nova.compute.provider_tree [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.065819] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Releasing lock "refresh_cache-650d2875-68f7-4131-b9b7-9799ee08f1d6" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.066188] env[61728]: DEBUG nova.compute.manager [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Instance network_info: |[{"id": "fdd5e1d2-60d3-4304-a31b-e6b680c01c63", "address": "fa:16:3e:1b:8d:97", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.126", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdd5e1d2-60", "ovs_interfaceid": "fdd5e1d2-60d3-4304-a31b-e6b680c01c63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 673.066779] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:8d:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '457c42cd-4ddb-4374-923e-d419b7f6eaff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fdd5e1d2-60d3-4304-a31b-e6b680c01c63', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 673.074829] env[61728]: DEBUG oslo.service.loopingcall [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 673.075090] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 673.075328] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4161e7e-3973-447f-9d61-f8555b8b4680 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.098237] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 673.098237] env[61728]: value = "task-463841" [ 673.098237] env[61728]: _type = "Task" [ 673.098237] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.113357] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463841, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.237999] env[61728]: DEBUG nova.compute.manager [req-f2206f04-6f62-44c5-9de6-0af504a0a2f3 req-3e40c1b5-d87a-46e1-93a6-50998fb0ea14 service nova] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Received event network-vif-plugged-fdd5e1d2-60d3-4304-a31b-e6b680c01c63 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 673.238257] env[61728]: DEBUG oslo_concurrency.lockutils [req-f2206f04-6f62-44c5-9de6-0af504a0a2f3 req-3e40c1b5-d87a-46e1-93a6-50998fb0ea14 service nova] Acquiring lock "650d2875-68f7-4131-b9b7-9799ee08f1d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.238468] env[61728]: DEBUG oslo_concurrency.lockutils [req-f2206f04-6f62-44c5-9de6-0af504a0a2f3 req-3e40c1b5-d87a-46e1-93a6-50998fb0ea14 service nova] Lock "650d2875-68f7-4131-b9b7-9799ee08f1d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.238638] env[61728]: DEBUG oslo_concurrency.lockutils [req-f2206f04-6f62-44c5-9de6-0af504a0a2f3 req-3e40c1b5-d87a-46e1-93a6-50998fb0ea14 service nova] Lock "650d2875-68f7-4131-b9b7-9799ee08f1d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.238805] env[61728]: DEBUG nova.compute.manager [req-f2206f04-6f62-44c5-9de6-0af504a0a2f3 req-3e40c1b5-d87a-46e1-93a6-50998fb0ea14 service nova] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] No waiting events found dispatching network-vif-plugged-fdd5e1d2-60d3-4304-a31b-e6b680c01c63 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 673.238969] env[61728]: WARNING nova.compute.manager [req-f2206f04-6f62-44c5-9de6-0af504a0a2f3 req-3e40c1b5-d87a-46e1-93a6-50998fb0ea14 service nova] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Received unexpected event network-vif-plugged-fdd5e1d2-60d3-4304-a31b-e6b680c01c63 for instance with vm_state building and task_state spawning. [ 673.239148] env[61728]: DEBUG nova.compute.manager [req-f2206f04-6f62-44c5-9de6-0af504a0a2f3 req-3e40c1b5-d87a-46e1-93a6-50998fb0ea14 service nova] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Received event network-changed-fdd5e1d2-60d3-4304-a31b-e6b680c01c63 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 673.239299] env[61728]: DEBUG nova.compute.manager [req-f2206f04-6f62-44c5-9de6-0af504a0a2f3 req-3e40c1b5-d87a-46e1-93a6-50998fb0ea14 service nova] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Refreshing instance network info cache due to event network-changed-fdd5e1d2-60d3-4304-a31b-e6b680c01c63. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 673.239584] env[61728]: DEBUG oslo_concurrency.lockutils [req-f2206f04-6f62-44c5-9de6-0af504a0a2f3 req-3e40c1b5-d87a-46e1-93a6-50998fb0ea14 service nova] Acquiring lock "refresh_cache-650d2875-68f7-4131-b9b7-9799ee08f1d6" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.239745] env[61728]: DEBUG oslo_concurrency.lockutils [req-f2206f04-6f62-44c5-9de6-0af504a0a2f3 req-3e40c1b5-d87a-46e1-93a6-50998fb0ea14 service nova] Acquired lock "refresh_cache-650d2875-68f7-4131-b9b7-9799ee08f1d6" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.239907] env[61728]: DEBUG nova.network.neutron [req-f2206f04-6f62-44c5-9de6-0af504a0a2f3 req-3e40c1b5-d87a-46e1-93a6-50998fb0ea14 service nova] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Refreshing network info cache for port fdd5e1d2-60d3-4304-a31b-e6b680c01c63 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 673.368362] env[61728]: DEBUG oslo_vmware.api [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.376350] env[61728]: DEBUG nova.network.neutron [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Updating instance_info_cache with network_info: [{"id": "b5b01c97-ed6d-41f8-98e3-9489d2e04f12", "address": "fa:16:3e:65:b6:d1", "network": {"id": "cdb3d002-e23a-4825-b2f0-1ce2d309c853", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-233804218", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.215", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5b01c97-ed", "ovs_interfaceid": "b5b01c97-ed6d-41f8-98e3-9489d2e04f12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "48764907-e12d-424f-8b5b-20c70e244c6a", "address": "fa:16:3e:72:67:6f", "network": {"id": "318189c7-0751-40d3-b344-d3e2eadd96cb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-828928406", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.70", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48764907-e1", "ovs_interfaceid": "48764907-e12d-424f-8b5b-20c70e244c6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b24101a4-7e85-4510-b0ce-1ebede43eb5c", "address": "fa:16:3e:53:e5:18", "network": {"id": "cdb3d002-e23a-4825-b2f0-1ce2d309c853", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-233804218", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.98", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb24101a4-7e", "ovs_interfaceid": "b24101a4-7e85-4510-b0ce-1ebede43eb5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.398186] env[61728]: DEBUG nova.compute.manager [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 673.421296] env[61728]: DEBUG nova.virt.hardware [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 673.421655] env[61728]: DEBUG nova.virt.hardware [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 673.421829] env[61728]: DEBUG nova.virt.hardware [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 673.423019] env[61728]: DEBUG nova.virt.hardware [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 673.423019] env[61728]: DEBUG nova.virt.hardware [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 673.423019] env[61728]: DEBUG nova.virt.hardware [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 673.423019] env[61728]: DEBUG nova.virt.hardware [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 673.423019] env[61728]: DEBUG nova.virt.hardware [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 673.423019] env[61728]: DEBUG nova.virt.hardware [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 673.423317] env[61728]: DEBUG nova.virt.hardware [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 673.423317] env[61728]: DEBUG nova.virt.hardware [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 673.424170] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e388861e-55b0-46cb-987b-818a90f67155 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.433435] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c096da-6c23-494c-a621-505a104c7538 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.510111] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463840, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.569961} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.510277] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 507b5aba-aec9-438b-92c0-5efa17959bc7/507b5aba-aec9-438b-92c0-5efa17959bc7.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 673.510448] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 673.510710] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8a16941e-8109-46ee-8551-72706d739687 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.516643] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 673.516643] env[61728]: value = "task-463842" [ 673.516643] env[61728]: _type = "Task" [ 673.516643] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.525965] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463842, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.538652] env[61728]: DEBUG nova.scheduler.client.report [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 673.608533] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463841, 'name': CreateVM_Task, 'duration_secs': 0.332528} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.608707] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 673.609441] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.609634] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.610019] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 673.610351] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3847273-16e6-4d3a-9539-17d293ff6585 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.614959] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 673.614959] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525eb252-1761-cbe7-f724-95ddfda0d441" [ 673.614959] env[61728]: _type = "Task" [ 673.614959] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.625062] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525eb252-1761-cbe7-f724-95ddfda0d441, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.739111] env[61728]: DEBUG nova.network.neutron [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Successfully updated port: 2537e303-47b7-4ed7-b8b9-7fd10ffc69d5 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 673.870095] env[61728]: DEBUG oslo_vmware.api [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.879641] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Releasing lock "refresh_cache-3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.879907] env[61728]: DEBUG nova.compute.manager [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Instance network_info: |[{"id": "b5b01c97-ed6d-41f8-98e3-9489d2e04f12", "address": "fa:16:3e:65:b6:d1", "network": {"id": "cdb3d002-e23a-4825-b2f0-1ce2d309c853", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-233804218", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.215", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5b01c97-ed", "ovs_interfaceid": "b5b01c97-ed6d-41f8-98e3-9489d2e04f12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "48764907-e12d-424f-8b5b-20c70e244c6a", "address": "fa:16:3e:72:67:6f", "network": {"id": "318189c7-0751-40d3-b344-d3e2eadd96cb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-828928406", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.70", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48764907-e1", "ovs_interfaceid": "48764907-e12d-424f-8b5b-20c70e244c6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b24101a4-7e85-4510-b0ce-1ebede43eb5c", "address": "fa:16:3e:53:e5:18", "network": {"id": "cdb3d002-e23a-4825-b2f0-1ce2d309c853", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-233804218", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.98", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb24101a4-7e", "ovs_interfaceid": "b24101a4-7e85-4510-b0ce-1ebede43eb5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 673.880555] env[61728]: DEBUG oslo_concurrency.lockutils [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] Acquired lock "refresh_cache-3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.880555] env[61728]: DEBUG nova.network.neutron [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Refreshing network info cache for port 48764907-e12d-424f-8b5b-20c70e244c6a {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 673.882766] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:b6:d1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b5b01c97-ed6d-41f8-98e3-9489d2e04f12', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:67:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f8442aa5-73db-4599-8564-b98a6ea26b9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '48764907-e12d-424f-8b5b-20c70e244c6a', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:e5:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b24101a4-7e85-4510-b0ce-1ebede43eb5c', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 673.894417] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Creating folder: Project (037e1941b7964a2b94d5b0f5c280956e). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 673.895555] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e89be5e4-db6f-43d2-a659-b9e398c57471 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.905759] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Created folder: Project (037e1941b7964a2b94d5b0f5c280956e) in parent group-v121913. [ 673.905989] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Creating folder: Instances. Parent ref: group-v122037. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 673.908124] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4322c9ab-8651-443f-bd82-f9c1859711fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.915937] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Created folder: Instances in parent group-v122037. [ 673.916212] env[61728]: DEBUG oslo.service.loopingcall [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 673.916410] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 673.916636] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4306833b-8697-4b6e-b9d3-4760ce2decc9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.944481] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 673.944481] env[61728]: value = "task-463845" [ 673.944481] env[61728]: _type = "Task" [ 673.944481] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.952139] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463845, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.986931] env[61728]: DEBUG nova.network.neutron [req-f2206f04-6f62-44c5-9de6-0af504a0a2f3 req-3e40c1b5-d87a-46e1-93a6-50998fb0ea14 service nova] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Updated VIF entry in instance network info cache for port fdd5e1d2-60d3-4304-a31b-e6b680c01c63. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 673.987293] env[61728]: DEBUG nova.network.neutron [req-f2206f04-6f62-44c5-9de6-0af504a0a2f3 req-3e40c1b5-d87a-46e1-93a6-50998fb0ea14 service nova] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Updating instance_info_cache with network_info: [{"id": "fdd5e1d2-60d3-4304-a31b-e6b680c01c63", "address": "fa:16:3e:1b:8d:97", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.126", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdd5e1d2-60", "ovs_interfaceid": "fdd5e1d2-60d3-4304-a31b-e6b680c01c63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.026449] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463842, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071054} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.026762] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 674.027580] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e38a90-f7c0-4ee2-88c3-8f376d6b0a36 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.050535] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] 507b5aba-aec9-438b-92c0-5efa17959bc7/507b5aba-aec9-438b-92c0-5efa17959bc7.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 674.051408] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.679s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.051947] env[61728]: DEBUG nova.compute.manager [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 674.055136] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b884a3c-0993-4200-be88-c8f94aa8c696 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.070116] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.851s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.070365] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.072453] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.231s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.073992] env[61728]: INFO nova.compute.claims [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 674.083676] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 674.083676] env[61728]: value = "task-463846" [ 674.083676] env[61728]: _type = "Task" [ 674.083676] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.092963] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463846, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.099033] env[61728]: INFO nova.scheduler.client.report [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Deleted allocations for instance 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4 [ 674.126335] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525eb252-1761-cbe7-f724-95ddfda0d441, 'name': SearchDatastore_Task, 'duration_secs': 0.011841} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.126646] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.126904] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 674.128035] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.128035] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.128035] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 674.128208] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2441c819-a07b-4648-b71d-6e3ecb1d6393 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.138337] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 674.138556] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 674.139896] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08668f10-98ca-4149-b394-00d78332a1eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.145068] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 674.145068] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527372e0-3a60-4873-1747-3892d5d53c73" [ 674.145068] env[61728]: _type = "Task" [ 674.145068] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.152846] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527372e0-3a60-4873-1747-3892d5d53c73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.242056] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Acquiring lock "refresh_cache-ad304c11-4ff8-4910-92a5-24216d727c4b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.242056] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Acquired lock "refresh_cache-ad304c11-4ff8-4910-92a5-24216d727c4b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.242247] env[61728]: DEBUG nova.network.neutron [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 674.370042] env[61728]: DEBUG oslo_vmware.api [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.455601] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463845, 'name': CreateVM_Task, 'duration_secs': 0.451311} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.458049] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 674.458911] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.459166] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.459497] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 674.460082] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-265c8314-e002-4f0e-97c7-d027eb62c387 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.464796] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 674.464796] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eaa2dd-07d9-38cd-2158-ec9e83da46fe" [ 674.464796] env[61728]: _type = "Task" [ 674.464796] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.473117] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eaa2dd-07d9-38cd-2158-ec9e83da46fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.490198] env[61728]: DEBUG oslo_concurrency.lockutils [req-f2206f04-6f62-44c5-9de6-0af504a0a2f3 req-3e40c1b5-d87a-46e1-93a6-50998fb0ea14 service nova] Releasing lock "refresh_cache-650d2875-68f7-4131-b9b7-9799ee08f1d6" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.557587] env[61728]: DEBUG nova.compute.utils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 674.559364] env[61728]: DEBUG nova.compute.manager [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 674.559364] env[61728]: DEBUG nova.network.neutron [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 674.596767] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463846, 'name': ReconfigVM_Task, 'duration_secs': 0.295888} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.597077] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Reconfigured VM instance instance-00000027 to attach disk [datastore1] 507b5aba-aec9-438b-92c0-5efa17959bc7/507b5aba-aec9-438b-92c0-5efa17959bc7.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 674.597732] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0ba5d5b8-7d1c-42e3-9cab-4b9b51ffd8e0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.606669] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 674.606669] env[61728]: value = "task-463847" [ 674.606669] env[61728]: _type = "Task" [ 674.606669] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.607113] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9f451b05-66f2-4cc3-9705-225975e18a6f tempest-DeleteServersAdminTestJSON-2109413549 tempest-DeleteServersAdminTestJSON-2109413549-project-member] Lock "7a60fcd0-01b0-40bc-93e7-6f5ce03232d4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.523s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.617665] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463847, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.626526] env[61728]: DEBUG nova.network.neutron [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Updated VIF entry in instance network info cache for port 48764907-e12d-424f-8b5b-20c70e244c6a. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 674.627011] env[61728]: DEBUG nova.network.neutron [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Updating instance_info_cache with network_info: [{"id": "b5b01c97-ed6d-41f8-98e3-9489d2e04f12", "address": "fa:16:3e:65:b6:d1", "network": {"id": "cdb3d002-e23a-4825-b2f0-1ce2d309c853", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-233804218", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.215", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5b01c97-ed", "ovs_interfaceid": "b5b01c97-ed6d-41f8-98e3-9489d2e04f12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "48764907-e12d-424f-8b5b-20c70e244c6a", "address": "fa:16:3e:72:67:6f", "network": {"id": "318189c7-0751-40d3-b344-d3e2eadd96cb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-828928406", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.70", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48764907-e1", "ovs_interfaceid": "48764907-e12d-424f-8b5b-20c70e244c6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b24101a4-7e85-4510-b0ce-1ebede43eb5c", "address": "fa:16:3e:53:e5:18", "network": {"id": "cdb3d002-e23a-4825-b2f0-1ce2d309c853", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-233804218", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.98", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb24101a4-7e", "ovs_interfaceid": "b24101a4-7e85-4510-b0ce-1ebede43eb5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.632403] env[61728]: DEBUG nova.policy [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '489b3e9916b8461a845e656c7ce2bd2f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cad2b09b986d4096ad91489a99b8d27e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 674.654834] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527372e0-3a60-4873-1747-3892d5d53c73, 'name': SearchDatastore_Task, 'duration_secs': 0.011502} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.655600] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2cba0cb-3ad0-4491-b133-cf170e0fc069 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.660695] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 674.660695] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5248dacb-f64a-b92a-0f07-2a8be1306b05" [ 674.660695] env[61728]: _type = "Task" [ 674.660695] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.668649] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5248dacb-f64a-b92a-0f07-2a8be1306b05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.696205] env[61728]: DEBUG oslo_vmware.rw_handles [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52cf9ca3-103c-0d0d-1b2a-df57638b0204/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 674.697326] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9917f927-2337-4a3f-a42f-806fc7bf5bfe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.703402] env[61728]: DEBUG oslo_vmware.rw_handles [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52cf9ca3-103c-0d0d-1b2a-df57638b0204/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 674.703568] env[61728]: ERROR oslo_vmware.rw_handles [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52cf9ca3-103c-0d0d-1b2a-df57638b0204/disk-0.vmdk due to incomplete transfer. [ 674.704158] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7528fb9c-e4a1-4739-9b7a-e4c39db6062b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.710599] env[61728]: DEBUG oslo_vmware.rw_handles [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52cf9ca3-103c-0d0d-1b2a-df57638b0204/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 674.710820] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Uploaded image b7db4964-4a64-4b56-bcc4-482c37cec071 to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 674.713242] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 674.713497] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-644be693-28e5-4cb8-a75c-1cddc75f1c1d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.718828] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 674.718828] env[61728]: value = "task-463848" [ 674.718828] env[61728]: _type = "Task" [ 674.718828] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.727951] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463848, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.785719] env[61728]: DEBUG nova.network.neutron [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.870558] env[61728]: DEBUG oslo_vmware.api [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.978011] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eaa2dd-07d9-38cd-2158-ec9e83da46fe, 'name': SearchDatastore_Task, 'duration_secs': 0.01238} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.981188] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.981644] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 674.982785] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.031936] env[61728]: DEBUG nova.network.neutron [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Successfully created port: a808f8cb-f136-434e-8bf7-5ee0b957e156 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 675.038825] env[61728]: DEBUG nova.network.neutron [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Updating instance_info_cache with network_info: [{"id": "2537e303-47b7-4ed7-b8b9-7fd10ffc69d5", "address": "fa:16:3e:8a:c7:f3", "network": {"id": "02755b4a-0ec6-4273-b3ee-9046ca905d78", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1247925051-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad57e3e090854cee813c4dfd8eea26ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2537e303-47", "ovs_interfaceid": "2537e303-47b7-4ed7-b8b9-7fd10ffc69d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.061721] env[61728]: DEBUG nova.compute.manager [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 675.119132] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463847, 'name': Rename_Task, 'duration_secs': 0.153218} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.121595] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 675.122291] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2b975b4-61b5-4042-9d1c-60efa06b6e2a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.128749] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 675.128749] env[61728]: value = "task-463849" [ 675.128749] env[61728]: _type = "Task" [ 675.128749] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.138123] env[61728]: DEBUG oslo_concurrency.lockutils [req-164a99a4-774c-4776-a6de-7793fea06341 req-d3adb0b0-0ce6-41f5-8daf-0de017a3b2c3 service nova] Releasing lock "refresh_cache-3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.138678] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463849, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.174163] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5248dacb-f64a-b92a-0f07-2a8be1306b05, 'name': SearchDatastore_Task, 'duration_secs': 0.009529} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.174383] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.174643] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 650d2875-68f7-4131-b9b7-9799ee08f1d6/650d2875-68f7-4131-b9b7-9799ee08f1d6.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 675.174921] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.175131] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 675.175332] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-99fccb37-91a7-4bea-bf14-3fa222cd8bb4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.178125] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6735575d-0d42-486b-96a3-f33284a37b57 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.184172] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 675.184172] env[61728]: value = "task-463850" [ 675.184172] env[61728]: _type = "Task" [ 675.184172] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.190420] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 675.190602] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 675.191763] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfb1f86a-49b6-4b9e-902a-f470b45863b2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.197138] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463850, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.200864] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 675.200864] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bb8cdd-2997-795d-2ce2-93c23f25412f" [ 675.200864] env[61728]: _type = "Task" [ 675.200864] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.211801] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bb8cdd-2997-795d-2ce2-93c23f25412f, 'name': SearchDatastore_Task, 'duration_secs': 0.00857} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.212627] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ef98b8c-a5b1-4392-aeb3-f1a45edf7161 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.219377] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 675.219377] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5239b3a9-10e7-68c9-1c93-704b6d2c144f" [ 675.219377] env[61728]: _type = "Task" [ 675.219377] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.233967] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5239b3a9-10e7-68c9-1c93-704b6d2c144f, 'name': SearchDatastore_Task, 'duration_secs': 0.008281} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.238360] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.238681] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1/3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 675.238987] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463848, 'name': Destroy_Task, 'duration_secs': 0.345775} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.242638] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-45c862d5-d35f-4e8d-862d-f3098acb459e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.245419] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Destroyed the VM [ 675.245701] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 675.246155] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-feb053da-b4d6-41e1-9a0a-dc008904ab3f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.252492] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 675.252492] env[61728]: value = "task-463851" [ 675.252492] env[61728]: _type = "Task" [ 675.252492] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.253879] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 675.253879] env[61728]: value = "task-463852" [ 675.253879] env[61728]: _type = "Task" [ 675.253879] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.266317] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463851, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.272245] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463852, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.373315] env[61728]: DEBUG oslo_vmware.api [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.420410] env[61728]: DEBUG nova.compute.manager [req-d03a4f46-f8a3-4106-b558-79aebcc88e19 req-22167592-86e4-4bd5-9e14-27120ba9c72a service nova] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Received event network-vif-plugged-2537e303-47b7-4ed7-b8b9-7fd10ffc69d5 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 675.420691] env[61728]: DEBUG oslo_concurrency.lockutils [req-d03a4f46-f8a3-4106-b558-79aebcc88e19 req-22167592-86e4-4bd5-9e14-27120ba9c72a service nova] Acquiring lock "ad304c11-4ff8-4910-92a5-24216d727c4b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.420970] env[61728]: DEBUG oslo_concurrency.lockutils [req-d03a4f46-f8a3-4106-b558-79aebcc88e19 req-22167592-86e4-4bd5-9e14-27120ba9c72a service nova] Lock "ad304c11-4ff8-4910-92a5-24216d727c4b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.421203] env[61728]: DEBUG oslo_concurrency.lockutils [req-d03a4f46-f8a3-4106-b558-79aebcc88e19 req-22167592-86e4-4bd5-9e14-27120ba9c72a service nova] Lock "ad304c11-4ff8-4910-92a5-24216d727c4b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.421412] env[61728]: DEBUG nova.compute.manager [req-d03a4f46-f8a3-4106-b558-79aebcc88e19 req-22167592-86e4-4bd5-9e14-27120ba9c72a service nova] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] No waiting events found dispatching network-vif-plugged-2537e303-47b7-4ed7-b8b9-7fd10ffc69d5 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 675.421872] env[61728]: WARNING nova.compute.manager [req-d03a4f46-f8a3-4106-b558-79aebcc88e19 req-22167592-86e4-4bd5-9e14-27120ba9c72a service nova] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Received unexpected event network-vif-plugged-2537e303-47b7-4ed7-b8b9-7fd10ffc69d5 for instance with vm_state building and task_state spawning. [ 675.422139] env[61728]: DEBUG nova.compute.manager [req-d03a4f46-f8a3-4106-b558-79aebcc88e19 req-22167592-86e4-4bd5-9e14-27120ba9c72a service nova] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Received event network-changed-2537e303-47b7-4ed7-b8b9-7fd10ffc69d5 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 675.422339] env[61728]: DEBUG nova.compute.manager [req-d03a4f46-f8a3-4106-b558-79aebcc88e19 req-22167592-86e4-4bd5-9e14-27120ba9c72a service nova] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Refreshing instance network info cache due to event network-changed-2537e303-47b7-4ed7-b8b9-7fd10ffc69d5. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 675.423394] env[61728]: DEBUG oslo_concurrency.lockutils [req-d03a4f46-f8a3-4106-b558-79aebcc88e19 req-22167592-86e4-4bd5-9e14-27120ba9c72a service nova] Acquiring lock "refresh_cache-ad304c11-4ff8-4910-92a5-24216d727c4b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.541298] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Releasing lock "refresh_cache-ad304c11-4ff8-4910-92a5-24216d727c4b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.541776] env[61728]: DEBUG nova.compute.manager [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Instance network_info: |[{"id": "2537e303-47b7-4ed7-b8b9-7fd10ffc69d5", "address": "fa:16:3e:8a:c7:f3", "network": {"id": "02755b4a-0ec6-4273-b3ee-9046ca905d78", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1247925051-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad57e3e090854cee813c4dfd8eea26ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2537e303-47", "ovs_interfaceid": "2537e303-47b7-4ed7-b8b9-7fd10ffc69d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 675.542205] env[61728]: DEBUG oslo_concurrency.lockutils [req-d03a4f46-f8a3-4106-b558-79aebcc88e19 req-22167592-86e4-4bd5-9e14-27120ba9c72a service nova] Acquired lock "refresh_cache-ad304c11-4ff8-4910-92a5-24216d727c4b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.542452] env[61728]: DEBUG nova.network.neutron [req-d03a4f46-f8a3-4106-b558-79aebcc88e19 req-22167592-86e4-4bd5-9e14-27120ba9c72a service nova] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Refreshing network info cache for port 2537e303-47b7-4ed7-b8b9-7fd10ffc69d5 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 675.543913] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:c7:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e044cfd4-1b0d-4d88-b1bd-604025731d3f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2537e303-47b7-4ed7-b8b9-7fd10ffc69d5', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 675.555099] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Creating folder: Project (ad57e3e090854cee813c4dfd8eea26ac). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 675.561859] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d80a378c-2b2d-4a25-89c1-c27f5a0026a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.581431] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Created folder: Project (ad57e3e090854cee813c4dfd8eea26ac) in parent group-v121913. [ 675.581797] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Creating folder: Instances. Parent ref: group-v122040. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 675.582857] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-38f3d35f-1032-4f1a-bbaf-c2c996166255 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.594266] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Created folder: Instances in parent group-v122040. [ 675.594556] env[61728]: DEBUG oslo.service.loopingcall [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 675.594762] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 675.595926] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eaf8600f-fe9c-48b0-bb03-c00374cca703 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.626792] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 675.626792] env[61728]: value = "task-463855" [ 675.626792] env[61728]: _type = "Task" [ 675.626792] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.646559] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463855, 'name': CreateVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.650589] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463849, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.700504] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463850, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472833} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.703590] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 650d2875-68f7-4131-b9b7-9799ee08f1d6/650d2875-68f7-4131-b9b7-9799ee08f1d6.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 675.703889] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 675.704264] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f68be1d4-0535-4445-8e28-571c677c34d4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.712831] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 675.712831] env[61728]: value = "task-463856" [ 675.712831] env[61728]: _type = "Task" [ 675.712831] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.727305] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463856, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.729658] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f936a978-bdf3-4947-bb49-50c7946495d9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.738231] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dae15e1-6dd7-49ad-b6cd-9d1a41155db6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.784022] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f88bce5-5beb-4480-ae9e-4be4936f3acc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.797866] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463851, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.803958] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463852, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.805804] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a70a1ae-bb15-48f8-bda9-fe5e6605440e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.822173] env[61728]: DEBUG nova.compute.provider_tree [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.877234] env[61728]: DEBUG oslo_vmware.api [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.877234] env[61728]: DEBUG nova.network.neutron [req-d03a4f46-f8a3-4106-b558-79aebcc88e19 req-22167592-86e4-4bd5-9e14-27120ba9c72a service nova] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Updated VIF entry in instance network info cache for port 2537e303-47b7-4ed7-b8b9-7fd10ffc69d5. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 675.877234] env[61728]: DEBUG nova.network.neutron [req-d03a4f46-f8a3-4106-b558-79aebcc88e19 req-22167592-86e4-4bd5-9e14-27120ba9c72a service nova] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Updating instance_info_cache with network_info: [{"id": "2537e303-47b7-4ed7-b8b9-7fd10ffc69d5", "address": "fa:16:3e:8a:c7:f3", "network": {"id": "02755b4a-0ec6-4273-b3ee-9046ca905d78", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1247925051-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad57e3e090854cee813c4dfd8eea26ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2537e303-47", "ovs_interfaceid": "2537e303-47b7-4ed7-b8b9-7fd10ffc69d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.073153] env[61728]: DEBUG nova.compute.manager [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 676.100543] env[61728]: DEBUG nova.virt.hardware [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 676.100828] env[61728]: DEBUG nova.virt.hardware [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 676.100996] env[61728]: DEBUG nova.virt.hardware [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 676.101228] env[61728]: DEBUG nova.virt.hardware [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 676.101470] env[61728]: DEBUG nova.virt.hardware [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 676.101549] env[61728]: DEBUG nova.virt.hardware [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 676.101799] env[61728]: DEBUG nova.virt.hardware [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 676.102060] env[61728]: DEBUG nova.virt.hardware [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 676.102202] env[61728]: DEBUG nova.virt.hardware [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 676.102367] env[61728]: DEBUG nova.virt.hardware [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 676.102552] env[61728]: DEBUG nova.virt.hardware [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 676.103539] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef7aa568-6b8b-4d61-ab68-d14ab20134f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.111951] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4f1017-8b19-48c2-89e2-806df7412067 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.145417] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463855, 'name': CreateVM_Task, 'duration_secs': 0.402838} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.149722] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 676.150467] env[61728]: DEBUG oslo_vmware.api [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463849, 'name': PowerOnVM_Task, 'duration_secs': 0.542443} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.151161] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.151324] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.151667] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 676.151893] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 676.152184] env[61728]: INFO nova.compute.manager [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Took 8.11 seconds to spawn the instance on the hypervisor. [ 676.152336] env[61728]: DEBUG nova.compute.manager [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 676.152589] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc2937b1-952a-4519-b5e2-0c72e1f04b4b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.154901] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a86a7b6-7b4f-414b-acf4-63d4ec6d6659 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.160358] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Waiting for the task: (returnval){ [ 676.160358] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52502d4a-0d8b-ae38-00f0-80c4fb546e37" [ 676.160358] env[61728]: _type = "Task" [ 676.160358] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.172821] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52502d4a-0d8b-ae38-00f0-80c4fb546e37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.222376] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463856, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087741} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.222839] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 676.225079] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4548a6-c7c3-4bdd-84b0-1f342540f3f8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.247744] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Reconfiguring VM instance instance-00000028 to attach disk [datastore1] 650d2875-68f7-4131-b9b7-9799ee08f1d6/650d2875-68f7-4131-b9b7-9799ee08f1d6.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 676.248090] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3cf403bf-20f5-40c9-93ae-6c95c86192d4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.268049] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 676.268049] env[61728]: value = "task-463857" [ 676.268049] env[61728]: _type = "Task" [ 676.268049] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.277247] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463857, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.289068] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463851, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.75095} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.292546] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1/3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 676.292837] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 676.293147] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463852, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.293374] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-33c0377f-4ea2-4a21-b6a9-1f86062c3b83 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.300707] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 676.300707] env[61728]: value = "task-463858" [ 676.300707] env[61728]: _type = "Task" [ 676.300707] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.310026] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463858, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.325523] env[61728]: DEBUG nova.scheduler.client.report [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.372901] env[61728]: DEBUG oslo_vmware.api [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463839, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.379580] env[61728]: DEBUG oslo_concurrency.lockutils [req-d03a4f46-f8a3-4106-b558-79aebcc88e19 req-22167592-86e4-4bd5-9e14-27120ba9c72a service nova] Releasing lock "refresh_cache-ad304c11-4ff8-4910-92a5-24216d727c4b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.584727] env[61728]: DEBUG nova.network.neutron [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Successfully updated port: a808f8cb-f136-434e-8bf7-5ee0b957e156 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 676.682031] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52502d4a-0d8b-ae38-00f0-80c4fb546e37, 'name': SearchDatastore_Task, 'duration_secs': 0.012915} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.683027] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.683345] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 676.683639] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.683788] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.683998] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 676.684561] env[61728]: INFO nova.compute.manager [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Took 50.71 seconds to build instance. [ 676.685438] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a09b85dc-d846-4bb6-aeec-faeb11f3bcd8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.694386] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 676.694590] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 676.695379] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de74b292-393f-41a0-b7d1-f36655146ee5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.701434] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Waiting for the task: (returnval){ [ 676.701434] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eeac7d-53f6-0a10-79ba-8a7d626f800a" [ 676.701434] env[61728]: _type = "Task" [ 676.701434] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.710779] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eeac7d-53f6-0a10-79ba-8a7d626f800a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.780596] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463857, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.788815] env[61728]: DEBUG oslo_vmware.api [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463852, 'name': RemoveSnapshot_Task, 'duration_secs': 1.067128} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.789128] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 676.789358] env[61728]: INFO nova.compute.manager [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Took 13.66 seconds to snapshot the instance on the hypervisor. [ 676.809900] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463858, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.301955} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.810213] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 676.811019] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215ff14f-abd5-40a1-888b-21f331a466e1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.837783] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1/3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 676.839264] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.767s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.839752] env[61728]: DEBUG nova.compute.manager [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 676.842114] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6afd0f4-e119-45b9-b001-12470b64ab56 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.856787] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 27.767s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.857352] env[61728]: DEBUG nova.objects.instance [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61728) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 676.866161] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 676.866161] env[61728]: value = "task-463859" [ 676.866161] env[61728]: _type = "Task" [ 676.866161] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.877585] env[61728]: DEBUG oslo_vmware.api [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463839, 'name': ReconfigVM_Task, 'duration_secs': 5.743732} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.880789] env[61728]: DEBUG oslo_concurrency.lockutils [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.881139] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Reconfigured VM to detach interface {{(pid=61728) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 676.883960] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463859, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.088145] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "refresh_cache-b833f691-e76e-4a2e-94a4-7594fadc3ba8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.088388] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "refresh_cache-b833f691-e76e-4a2e-94a4-7594fadc3ba8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.088428] env[61728]: DEBUG nova.network.neutron [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 677.189080] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3b0118f5-ac2e-40b1-b09f-41dfa0adf470 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "507b5aba-aec9-438b-92c0-5efa17959bc7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.511s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.213265] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eeac7d-53f6-0a10-79ba-8a7d626f800a, 'name': SearchDatastore_Task, 'duration_secs': 0.008624} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.214321] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bde3dd21-1364-4a4e-9052-43157b5ef66d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.220082] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Waiting for the task: (returnval){ [ 677.220082] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52beb326-c4d3-6308-a5a4-fa6eb11380ce" [ 677.220082] env[61728]: _type = "Task" [ 677.220082] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.230847] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52beb326-c4d3-6308-a5a4-fa6eb11380ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.281155] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463857, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.359978] env[61728]: DEBUG nova.compute.manager [None req-75d67cc0-26c4-4e8e-aa4e-17758610892f tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Found 2 images (rotation: 2) {{(pid=61728) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 677.365272] env[61728]: DEBUG nova.compute.utils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 677.366522] env[61728]: DEBUG nova.compute.manager [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 677.366731] env[61728]: DEBUG nova.network.neutron [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 677.381800] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463859, 'name': ReconfigVM_Task, 'duration_secs': 0.290455} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.383950] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Reconfigured VM instance instance-00000026 to attach disk [datastore1] 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1/3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 677.385537] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b7eb24b2-20ac-412c-85aa-cb1dbfa304a2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.394390] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 677.394390] env[61728]: value = "task-463860" [ 677.394390] env[61728]: _type = "Task" [ 677.394390] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.409741] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463860, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.436218] env[61728]: DEBUG nova.policy [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '244030ae678b4d8588086675b6d23718', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ac86255b8224fc095ca68631066afbe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 677.508481] env[61728]: DEBUG nova.compute.manager [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Received event network-vif-plugged-a808f8cb-f136-434e-8bf7-5ee0b957e156 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 677.508856] env[61728]: DEBUG oslo_concurrency.lockutils [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] Acquiring lock "b833f691-e76e-4a2e-94a4-7594fadc3ba8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.508856] env[61728]: DEBUG oslo_concurrency.lockutils [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] Lock "b833f691-e76e-4a2e-94a4-7594fadc3ba8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.509245] env[61728]: DEBUG oslo_concurrency.lockutils [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] Lock "b833f691-e76e-4a2e-94a4-7594fadc3ba8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.509963] env[61728]: DEBUG nova.compute.manager [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] No waiting events found dispatching network-vif-plugged-a808f8cb-f136-434e-8bf7-5ee0b957e156 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 677.509963] env[61728]: WARNING nova.compute.manager [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Received unexpected event network-vif-plugged-a808f8cb-f136-434e-8bf7-5ee0b957e156 for instance with vm_state building and task_state spawning. [ 677.509963] env[61728]: DEBUG nova.compute.manager [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Received event network-changed-a808f8cb-f136-434e-8bf7-5ee0b957e156 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 677.510125] env[61728]: DEBUG nova.compute.manager [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Refreshing instance network info cache due to event network-changed-a808f8cb-f136-434e-8bf7-5ee0b957e156. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 677.510308] env[61728]: DEBUG oslo_concurrency.lockutils [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] Acquiring lock "refresh_cache-b833f691-e76e-4a2e-94a4-7594fadc3ba8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.634788] env[61728]: DEBUG nova.network.neutron [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.691864] env[61728]: DEBUG nova.compute.manager [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 677.738267] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52beb326-c4d3-6308-a5a4-fa6eb11380ce, 'name': SearchDatastore_Task, 'duration_secs': 0.028863} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.738267] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.738267] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] ad304c11-4ff8-4910-92a5-24216d727c4b/ad304c11-4ff8-4910-92a5-24216d727c4b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 677.738267] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c7baa2c-6632-4623-aedb-b02ee13f91f8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.744991] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Waiting for the task: (returnval){ [ 677.744991] env[61728]: value = "task-463861" [ 677.744991] env[61728]: _type = "Task" [ 677.744991] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.757372] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463861, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.785686] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463857, 'name': ReconfigVM_Task, 'duration_secs': 1.150615} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.785686] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Reconfigured VM instance instance-00000028 to attach disk [datastore1] 650d2875-68f7-4131-b9b7-9799ee08f1d6/650d2875-68f7-4131-b9b7-9799ee08f1d6.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 677.785686] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-701890af-c2b3-40bd-9363-e393b91e7771 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.792938] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 677.792938] env[61728]: value = "task-463862" [ 677.792938] env[61728]: _type = "Task" [ 677.792938] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.800033] env[61728]: DEBUG nova.network.neutron [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Successfully created port: 4d70757b-d260-4aa3-8914-a38eb567f7b8 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 677.806130] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463862, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.866908] env[61728]: DEBUG nova.compute.manager [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 677.874212] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa44907-5cd6-4d39-a9ee-fa64e6a082ba tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.876017] env[61728]: DEBUG nova.network.neutron [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Updating instance_info_cache with network_info: [{"id": "a808f8cb-f136-434e-8bf7-5ee0b957e156", "address": "fa:16:3e:63:52:85", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa808f8cb-f1", "ovs_interfaceid": "a808f8cb-f136-434e-8bf7-5ee0b957e156", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.877188] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.716s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.877402] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.879558] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.709s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.880057] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.882169] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.536s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.885042] env[61728]: INFO nova.compute.claims [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 677.911100] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463860, 'name': Rename_Task, 'duration_secs': 0.142687} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.912242] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 677.912678] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eaa33b2e-0033-481e-81fc-c7846a46fe64 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.924506] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 677.924506] env[61728]: value = "task-463863" [ 677.924506] env[61728]: _type = "Task" [ 677.924506] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.932296] env[61728]: INFO nova.scheduler.client.report [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Deleted allocations for instance 26220c0c-f535-4315-b3bb-2e5ac4d2286e [ 677.934652] env[61728]: INFO nova.scheduler.client.report [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Deleted allocations for instance fbd521c8-9618-4b5c-839e-50d8631fa7b9 [ 677.946364] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463863, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.040769] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "507b5aba-aec9-438b-92c0-5efa17959bc7" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.041084] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "507b5aba-aec9-438b-92c0-5efa17959bc7" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.211814] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.255788] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463861, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47176} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.255998] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] ad304c11-4ff8-4910-92a5-24216d727c4b/ad304c11-4ff8-4910-92a5-24216d727c4b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 678.256251] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 678.256512] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-00e44926-4125-4b42-9353-8fc1670e9bbe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.262671] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Waiting for the task: (returnval){ [ 678.262671] env[61728]: value = "task-463864" [ 678.262671] env[61728]: _type = "Task" [ 678.262671] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.270323] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463864, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.302072] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463862, 'name': Rename_Task, 'duration_secs': 0.153239} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.302730] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 678.302986] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ab20cc0-b4fd-4182-b52f-d066bd3c7931 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.308346] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 678.308346] env[61728]: value = "task-463865" [ 678.308346] env[61728]: _type = "Task" [ 678.308346] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.316013] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463865, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.319620] env[61728]: DEBUG oslo_concurrency.lockutils [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.319844] env[61728]: DEBUG oslo_concurrency.lockutils [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.320095] env[61728]: DEBUG nova.network.neutron [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 678.387305] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "refresh_cache-b833f691-e76e-4a2e-94a4-7594fadc3ba8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.388358] env[61728]: DEBUG nova.compute.manager [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Instance network_info: |[{"id": "a808f8cb-f136-434e-8bf7-5ee0b957e156", "address": "fa:16:3e:63:52:85", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa808f8cb-f1", "ovs_interfaceid": "a808f8cb-f136-434e-8bf7-5ee0b957e156", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 678.390172] env[61728]: DEBUG oslo_concurrency.lockutils [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] Acquired lock "refresh_cache-b833f691-e76e-4a2e-94a4-7594fadc3ba8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.390396] env[61728]: DEBUG nova.network.neutron [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Refreshing network info cache for port a808f8cb-f136-434e-8bf7-5ee0b957e156 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 678.391698] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:52:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f65996a3-f865-4492-9377-cd14ec8b3aae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a808f8cb-f136-434e-8bf7-5ee0b957e156', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 678.400518] env[61728]: DEBUG oslo.service.loopingcall [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 678.402390] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 678.402618] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a52107d-ad16-4afb-b2db-d1d61889c033 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.423457] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 678.423457] env[61728]: value = "task-463866" [ 678.423457] env[61728]: _type = "Task" [ 678.423457] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.435544] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463866, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.438890] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463863, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.446661] env[61728]: DEBUG oslo_concurrency.lockutils [None req-07308936-2d04-4251-bb9a-9118ee5c99fd tempest-ListServerFiltersTestJSON-2139120249 tempest-ListServerFiltersTestJSON-2139120249-project-member] Lock "26220c0c-f535-4315-b3bb-2e5ac4d2286e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.974s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.451022] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa37d545-b59d-4a94-a3fa-099724dd33fe tempest-FloatingIPsAssociationNegativeTestJSON-672220382 tempest-FloatingIPsAssociationNegativeTestJSON-672220382-project-member] Lock "fbd521c8-9618-4b5c-839e-50d8631fa7b9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.530s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.544821] env[61728]: DEBUG nova.compute.utils [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 678.776020] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463864, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059616} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.776538] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 678.777403] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e3d7c6-7cf7-43b9-8d8f-6e069ede34be {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.802548] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] ad304c11-4ff8-4910-92a5-24216d727c4b/ad304c11-4ff8-4910-92a5-24216d727c4b.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 678.803643] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a46eb78-9f09-4717-ae9d-cb072178f02f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.827880] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Waiting for the task: (returnval){ [ 678.827880] env[61728]: value = "task-463867" [ 678.827880] env[61728]: _type = "Task" [ 678.827880] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.831306] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463865, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.840086] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463867, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.879285] env[61728]: DEBUG nova.compute.manager [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 678.909232] env[61728]: DEBUG nova.virt.hardware [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:24:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='ec856338-a71e-43d6-aa12-96f76d31f40f',id=36,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1304387059',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 678.909580] env[61728]: DEBUG nova.virt.hardware [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 678.909801] env[61728]: DEBUG nova.virt.hardware [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 678.910080] env[61728]: DEBUG nova.virt.hardware [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 678.910260] env[61728]: DEBUG nova.virt.hardware [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 678.910458] env[61728]: DEBUG nova.virt.hardware [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 678.910714] env[61728]: DEBUG nova.virt.hardware [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 678.910970] env[61728]: DEBUG nova.virt.hardware [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 678.911149] env[61728]: DEBUG nova.virt.hardware [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 678.911366] env[61728]: DEBUG nova.virt.hardware [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 678.911600] env[61728]: DEBUG nova.virt.hardware [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 678.912856] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70ca4f6-781c-43e3-a930-e9b57ab70110 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.928547] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9395eee-3bd6-4805-8a18-8f1692cd1580 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.952991] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463866, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.956674] env[61728]: DEBUG oslo_vmware.api [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463863, 'name': PowerOnVM_Task, 'duration_secs': 0.55116} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.956957] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 678.957179] env[61728]: INFO nova.compute.manager [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Took 13.60 seconds to spawn the instance on the hypervisor. [ 678.957362] env[61728]: DEBUG nova.compute.manager [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 678.958226] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42909af-80f4-4226-8e30-c66077b08979 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.048196] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "507b5aba-aec9-438b-92c0-5efa17959bc7" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.172828] env[61728]: DEBUG nova.compute.manager [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 679.173589] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2ebde9-96a5-4daa-9610-217c17237b26 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.226254] env[61728]: INFO nova.network.neutron [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Port f605e2a3-ef9e-4642-8d18-e43436149349 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 679.226254] env[61728]: DEBUG nova.network.neutron [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Updating instance_info_cache with network_info: [{"id": "edc90742-f02d-42ad-85cd-b4242f797c33", "address": "fa:16:3e:aa:75:a5", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedc90742-f0", "ovs_interfaceid": "edc90742-f02d-42ad-85cd-b4242f797c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.277269] env[61728]: DEBUG oslo_concurrency.lockutils [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "620bb472-c36e-4c56-acdf-42e02614856b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.277527] env[61728]: DEBUG oslo_concurrency.lockutils [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "620bb472-c36e-4c56-acdf-42e02614856b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.278191] env[61728]: DEBUG oslo_concurrency.lockutils [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "620bb472-c36e-4c56-acdf-42e02614856b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.278392] env[61728]: DEBUG oslo_concurrency.lockutils [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "620bb472-c36e-4c56-acdf-42e02614856b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.278574] env[61728]: DEBUG oslo_concurrency.lockutils [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "620bb472-c36e-4c56-acdf-42e02614856b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.281219] env[61728]: INFO nova.compute.manager [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Terminating instance [ 679.283813] env[61728]: DEBUG nova.compute.manager [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 679.283813] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 679.284929] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d921238c-d7ec-4619-825b-56633b114704 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.292926] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 679.293259] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af5ce4ea-94cb-4a53-b934-70118091ebde {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.295245] env[61728]: DEBUG nova.network.neutron [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Updated VIF entry in instance network info cache for port a808f8cb-f136-434e-8bf7-5ee0b957e156. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 679.295582] env[61728]: DEBUG nova.network.neutron [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Updating instance_info_cache with network_info: [{"id": "a808f8cb-f136-434e-8bf7-5ee0b957e156", "address": "fa:16:3e:63:52:85", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa808f8cb-f1", "ovs_interfaceid": "a808f8cb-f136-434e-8bf7-5ee0b957e156", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.301449] env[61728]: DEBUG oslo_vmware.api [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 679.301449] env[61728]: value = "task-463868" [ 679.301449] env[61728]: _type = "Task" [ 679.301449] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.314265] env[61728]: DEBUG oslo_vmware.api [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463868, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.330138] env[61728]: DEBUG oslo_vmware.api [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463865, 'name': PowerOnVM_Task, 'duration_secs': 0.921597} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.333017] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 679.333240] env[61728]: INFO nova.compute.manager [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Took 8.60 seconds to spawn the instance on the hypervisor. [ 679.333418] env[61728]: DEBUG nova.compute.manager [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 679.334614] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478299a2-dc08-4f51-a83a-f65e87f5976c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.349887] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463867, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.435991] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463866, 'name': CreateVM_Task, 'duration_secs': 0.637355} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.436184] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 679.436865] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.437049] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.437371] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 679.437623] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-803aa84a-4d8e-4a33-843a-eba2252c634d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.445228] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 679.445228] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52332ebb-fe9f-2a4b-936f-85002f8946c6" [ 679.445228] env[61728]: _type = "Task" [ 679.445228] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.452274] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52332ebb-fe9f-2a4b-936f-85002f8946c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.459729] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d6ea25d-8a12-482d-bee6-93b3e1a07020 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.466546] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c354b9-fc8c-467e-a88a-3db9b3f97210 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.498998] env[61728]: DEBUG nova.network.neutron [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Successfully updated port: 4d70757b-d260-4aa3-8914-a38eb567f7b8 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 679.508989] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aed53cb-0736-4387-90e2-8d6cc8dcc0b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.512255] env[61728]: INFO nova.compute.manager [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Took 55.81 seconds to build instance. [ 679.519079] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed6fac6-9c7a-454e-bd22-37b4dafb78cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.536370] env[61728]: DEBUG nova.compute.provider_tree [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 679.695328] env[61728]: INFO nova.compute.manager [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] instance snapshotting [ 679.695945] env[61728]: DEBUG nova.objects.instance [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'flavor' on Instance uuid 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 679.730223] env[61728]: DEBUG oslo_concurrency.lockutils [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "refresh_cache-620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.798540] env[61728]: DEBUG oslo_concurrency.lockutils [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] Releasing lock "refresh_cache-b833f691-e76e-4a2e-94a4-7594fadc3ba8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.798768] env[61728]: DEBUG nova.compute.manager [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Received event network-vif-deleted-f605e2a3-ef9e-4642-8d18-e43436149349 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 679.798947] env[61728]: INFO nova.compute.manager [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Neutron deleted interface f605e2a3-ef9e-4642-8d18-e43436149349; detaching it from the instance and deleting it from the info cache [ 679.799239] env[61728]: DEBUG nova.network.neutron [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Updating instance_info_cache with network_info: [{"id": "edc90742-f02d-42ad-85cd-b4242f797c33", "address": "fa:16:3e:aa:75:a5", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedc90742-f0", "ovs_interfaceid": "edc90742-f02d-42ad-85cd-b4242f797c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.815193] env[61728]: DEBUG oslo_vmware.api [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463868, 'name': PowerOffVM_Task, 'duration_secs': 0.243899} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.815833] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 679.815833] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 679.815989] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-69b67152-aa13-491d-96b3-b92e22ec5acd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.843438] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463867, 'name': ReconfigVM_Task, 'duration_secs': 0.77449} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.843723] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Reconfigured VM instance instance-00000029 to attach disk [datastore1] ad304c11-4ff8-4910-92a5-24216d727c4b/ad304c11-4ff8-4910-92a5-24216d727c4b.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 679.844401] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-01560e2a-a528-4090-953d-34d452c3fecd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.850977] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Waiting for the task: (returnval){ [ 679.850977] env[61728]: value = "task-463870" [ 679.850977] env[61728]: _type = "Task" [ 679.850977] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.865969] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463870, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.865969] env[61728]: INFO nova.compute.manager [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Took 46.70 seconds to build instance. [ 679.880632] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 679.880860] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 679.881095] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Deleting the datastore file [datastore1] 620bb472-c36e-4c56-acdf-42e02614856b {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 679.881420] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-afe718f5-b664-47f7-abfd-55f11fdefc3f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.887857] env[61728]: DEBUG oslo_vmware.api [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 679.887857] env[61728]: value = "task-463871" [ 679.887857] env[61728]: _type = "Task" [ 679.887857] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.897163] env[61728]: DEBUG oslo_vmware.api [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463871, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.956685] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52332ebb-fe9f-2a4b-936f-85002f8946c6, 'name': SearchDatastore_Task, 'duration_secs': 0.010388} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.957293] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.957720] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 679.958107] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.958410] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.958838] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 679.959250] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ee8e548f-859f-4b42-a01e-4dc790f34d19 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.969952] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 679.969952] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 679.972431] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-340cb3d8-5a0e-41db-b77c-f971bac416f8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.983326] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 679.983326] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5219fbe3-a41f-0d2c-8234-a9f3d14da6b3" [ 679.983326] env[61728]: _type = "Task" [ 679.983326] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.991084] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5219fbe3-a41f-0d2c-8234-a9f3d14da6b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.993857] env[61728]: DEBUG nova.compute.manager [req-4b32ac6a-c631-4ee6-bb1e-c1b310316674 req-74da3700-bc78-4be8-b84d-7aef36031bad service nova] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Received event network-vif-plugged-4d70757b-d260-4aa3-8914-a38eb567f7b8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 679.995244] env[61728]: DEBUG oslo_concurrency.lockutils [req-4b32ac6a-c631-4ee6-bb1e-c1b310316674 req-74da3700-bc78-4be8-b84d-7aef36031bad service nova] Acquiring lock "1d9114cb-a0af-4874-962f-27237b3c89cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.995244] env[61728]: DEBUG oslo_concurrency.lockutils [req-4b32ac6a-c631-4ee6-bb1e-c1b310316674 req-74da3700-bc78-4be8-b84d-7aef36031bad service nova] Lock "1d9114cb-a0af-4874-962f-27237b3c89cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.995393] env[61728]: DEBUG oslo_concurrency.lockutils [req-4b32ac6a-c631-4ee6-bb1e-c1b310316674 req-74da3700-bc78-4be8-b84d-7aef36031bad service nova] Lock "1d9114cb-a0af-4874-962f-27237b3c89cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.995825] env[61728]: DEBUG nova.compute.manager [req-4b32ac6a-c631-4ee6-bb1e-c1b310316674 req-74da3700-bc78-4be8-b84d-7aef36031bad service nova] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] No waiting events found dispatching network-vif-plugged-4d70757b-d260-4aa3-8914-a38eb567f7b8 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 679.996200] env[61728]: WARNING nova.compute.manager [req-4b32ac6a-c631-4ee6-bb1e-c1b310316674 req-74da3700-bc78-4be8-b84d-7aef36031bad service nova] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Received unexpected event network-vif-plugged-4d70757b-d260-4aa3-8914-a38eb567f7b8 for instance with vm_state building and task_state spawning. [ 679.996518] env[61728]: DEBUG nova.compute.manager [req-4b32ac6a-c631-4ee6-bb1e-c1b310316674 req-74da3700-bc78-4be8-b84d-7aef36031bad service nova] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Received event network-changed-4d70757b-d260-4aa3-8914-a38eb567f7b8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 679.996839] env[61728]: DEBUG nova.compute.manager [req-4b32ac6a-c631-4ee6-bb1e-c1b310316674 req-74da3700-bc78-4be8-b84d-7aef36031bad service nova] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Refreshing instance network info cache due to event network-changed-4d70757b-d260-4aa3-8914-a38eb567f7b8. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 679.998101] env[61728]: DEBUG oslo_concurrency.lockutils [req-4b32ac6a-c631-4ee6-bb1e-c1b310316674 req-74da3700-bc78-4be8-b84d-7aef36031bad service nova] Acquiring lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.998101] env[61728]: DEBUG oslo_concurrency.lockutils [req-4b32ac6a-c631-4ee6-bb1e-c1b310316674 req-74da3700-bc78-4be8-b84d-7aef36031bad service nova] Acquired lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.998101] env[61728]: DEBUG nova.network.neutron [req-4b32ac6a-c631-4ee6-bb1e-c1b310316674 req-74da3700-bc78-4be8-b84d-7aef36031bad service nova] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Refreshing network info cache for port 4d70757b-d260-4aa3-8914-a38eb567f7b8 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 680.009105] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.015401] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d500e77-ce91-41aa-83c4-5919e0d32acd tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.390s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.062872] env[61728]: ERROR nova.scheduler.client.report [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [req-3ed0035a-1932-4509-80b6-dceb93494033] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3ed0035a-1932-4509-80b6-dceb93494033"}]} [ 680.080678] env[61728]: DEBUG nova.scheduler.client.report [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 680.100451] env[61728]: DEBUG nova.scheduler.client.report [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 680.100451] env[61728]: DEBUG nova.compute.provider_tree [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 680.112207] env[61728]: DEBUG nova.scheduler.client.report [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 680.116819] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "507b5aba-aec9-438b-92c0-5efa17959bc7" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.117107] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "507b5aba-aec9-438b-92c0-5efa17959bc7" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.117371] env[61728]: INFO nova.compute.manager [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Attaching volume 10d75c82-d308-41da-a8bd-8beaea48adf9 to /dev/sdb [ 680.130941] env[61728]: DEBUG nova.scheduler.client.report [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 680.158833] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca4513b-3b20-4c87-bd64-9fe3e2ae6ac3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.174094] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5875d69-cfff-4c5a-be05-df419c33d6b1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.189459] env[61728]: DEBUG nova.virt.block_device [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Updating existing volume attachment record: f0218b29-71ef-4c27-b208-759bb0ac951e {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 680.206444] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05511167-bf8c-4a7f-9a2a-4e0817bc1aed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.228101] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95663aa4-e6d9-4aff-92a9-3d546d819de4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.236405] env[61728]: DEBUG oslo_concurrency.lockutils [None req-827d5fe4-df8c-43d5-a8b2-1ef893e46459 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-620bb472-c36e-4c56-acdf-42e02614856b-f605e2a3-ef9e-4642-8d18-e43436149349" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.977s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.305716] env[61728]: DEBUG oslo_concurrency.lockutils [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] Acquiring lock "620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.366702] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463870, 'name': Rename_Task, 'duration_secs': 0.195567} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.367920] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 680.367920] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d7d03982-2ee9-42d9-9a69-ac88e25197be {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.370134] env[61728]: DEBUG oslo_concurrency.lockutils [None req-71042408-53b5-4f31-b609-bd63bd8cbe82 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "650d2875-68f7-4131-b9b7-9799ee08f1d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.927s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.377655] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Waiting for the task: (returnval){ [ 680.377655] env[61728]: value = "task-463873" [ 680.377655] env[61728]: _type = "Task" [ 680.377655] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.387971] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463873, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.407105] env[61728]: DEBUG oslo_vmware.api [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-463871, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154374} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.407105] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 680.407105] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 680.407105] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 680.407105] env[61728]: INFO nova.compute.manager [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 680.407105] env[61728]: DEBUG oslo.service.loopingcall [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 680.407105] env[61728]: DEBUG nova.compute.manager [-] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.407105] env[61728]: DEBUG nova.network.neutron [-] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 680.496151] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5219fbe3-a41f-0d2c-8234-a9f3d14da6b3, 'name': SearchDatastore_Task, 'duration_secs': 0.00888} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.500030] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e81aa595-00fe-442a-968c-594c87f97f70 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.514162] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 680.514162] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52180831-6a16-fea2-e9ec-9818c78f6793" [ 680.514162] env[61728]: _type = "Task" [ 680.514162] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.517082] env[61728]: DEBUG nova.compute.manager [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 680.535925] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52180831-6a16-fea2-e9ec-9818c78f6793, 'name': SearchDatastore_Task, 'duration_secs': 0.013507} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.539731] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.540952] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b833f691-e76e-4a2e-94a4-7594fadc3ba8/b833f691-e76e-4a2e-94a4-7594fadc3ba8.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 680.540952] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5f6eabef-46f2-4ed3-9e4d-e0e88f910518 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.551417] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 680.551417] env[61728]: value = "task-463876" [ 680.551417] env[61728]: _type = "Task" [ 680.551417] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.562885] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463876, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.616016] env[61728]: DEBUG nova.network.neutron [req-4b32ac6a-c631-4ee6-bb1e-c1b310316674 req-74da3700-bc78-4be8-b84d-7aef36031bad service nova] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.673298] env[61728]: DEBUG oslo_concurrency.lockutils [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.673298] env[61728]: DEBUG oslo_concurrency.lockutils [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.673298] env[61728]: DEBUG oslo_concurrency.lockutils [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.673298] env[61728]: DEBUG oslo_concurrency.lockutils [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.673298] env[61728]: DEBUG oslo_concurrency.lockutils [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.674899] env[61728]: INFO nova.compute.manager [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Terminating instance [ 680.677957] env[61728]: DEBUG nova.compute.manager [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 680.677957] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 680.678458] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba8f0be8-2329-4f11-8d4f-fa8e61ed4856 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.692634] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 680.692946] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd435bc5-32a5-4667-ae5f-d02208d65e16 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.702068] env[61728]: DEBUG oslo_vmware.api [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 680.702068] env[61728]: value = "task-463877" [ 680.702068] env[61728]: _type = "Task" [ 680.702068] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.712407] env[61728]: DEBUG oslo_vmware.api [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463877, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.747055] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 680.747423] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-abfc8233-beb8-4a8b-a740-d28490de1b88 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.762070] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 680.762070] env[61728]: value = "task-463878" [ 680.762070] env[61728]: _type = "Task" [ 680.762070] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.772313] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463878, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.806766] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33ce52d-8a33-4628-8763-87d18c658af7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.814996] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa4ed56-f8fd-43e9-930b-c6f93a709aae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.852479] env[61728]: DEBUG nova.network.neutron [req-4b32ac6a-c631-4ee6-bb1e-c1b310316674 req-74da3700-bc78-4be8-b84d-7aef36031bad service nova] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.857340] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667de9cf-5eb9-4403-92c9-80913064fe1c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.864234] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb072f4-8ddb-4338-9cc8-bfdd8efd55a8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.639777] env[61728]: DEBUG nova.compute.manager [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 681.648572] env[61728]: DEBUG oslo_concurrency.lockutils [req-4b32ac6a-c631-4ee6-bb1e-c1b310316674 req-74da3700-bc78-4be8-b84d-7aef36031bad service nova] Releasing lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.661071] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.661273] env[61728]: DEBUG nova.network.neutron [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 681.675303] env[61728]: DEBUG nova.compute.provider_tree [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 681.687264] env[61728]: DEBUG oslo_vmware.api [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463873, 'name': PowerOnVM_Task, 'duration_secs': 0.879636} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.687344] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463876, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.540833} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.687532] env[61728]: DEBUG oslo_vmware.api [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463877, 'name': PowerOffVM_Task, 'duration_secs': 0.219268} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.688509] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.692302] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 681.692302] env[61728]: INFO nova.compute.manager [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Took 8.29 seconds to spawn the instance on the hypervisor. [ 681.692302] env[61728]: DEBUG nova.compute.manager [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 681.692518] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b833f691-e76e-4a2e-94a4-7594fadc3ba8/b833f691-e76e-4a2e-94a4-7594fadc3ba8.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 681.692765] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 681.693048] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 681.693217] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 681.697670] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa42e6f-4f5f-40ba-93cb-23fac86df18b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.702151] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-11f8e0c4-1fcf-4a23-8ad7-d97c2e241cbc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.702415] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6367060-664a-4fb4-a58b-bf2143a620f6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.704308] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463878, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.712241] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 681.712241] env[61728]: value = "task-463879" [ 681.712241] env[61728]: _type = "Task" [ 681.712241] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.718776] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d48a8f1d-92ae-4661-bd36-e17dc738cc0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.724149] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463879, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.727674] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-47ff5792-37f3-4772-96dd-d71f681a0954 tempest-ServersAdminNegativeTestJSON-1278040523 tempest-ServersAdminNegativeTestJSON-1278040523-project-admin] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Suspending the VM {{(pid=61728) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 681.727948] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-ae47aeaa-83cb-4ee6-a98a-40fd3b38d855 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.733378] env[61728]: DEBUG oslo_vmware.api [None req-47ff5792-37f3-4772-96dd-d71f681a0954 tempest-ServersAdminNegativeTestJSON-1278040523 tempest-ServersAdminNegativeTestJSON-1278040523-project-admin] Waiting for the task: (returnval){ [ 681.733378] env[61728]: value = "task-463881" [ 681.733378] env[61728]: _type = "Task" [ 681.733378] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.742482] env[61728]: DEBUG oslo_vmware.api [None req-47ff5792-37f3-4772-96dd-d71f681a0954 tempest-ServersAdminNegativeTestJSON-1278040523 tempest-ServersAdminNegativeTestJSON-1278040523-project-admin] Task: {'id': task-463881, 'name': SuspendVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.746822] env[61728]: DEBUG nova.scheduler.client.report [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 70 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 681.747047] env[61728]: DEBUG nova.compute.provider_tree [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 70 to 71 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 681.747250] env[61728]: DEBUG nova.compute.provider_tree [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 681.854902] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 681.857573] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 681.857573] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Deleting the datastore file [datastore1] 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 681.857573] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e792bb0a-e416-4b65-b559-d899a79fbe77 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.864526] env[61728]: DEBUG oslo_vmware.api [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 681.864526] env[61728]: value = "task-463882" [ 681.864526] env[61728]: _type = "Task" [ 681.864526] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.875198] env[61728]: DEBUG oslo_vmware.api [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463882, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.155633] env[61728]: DEBUG nova.network.neutron [-] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.164532] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463878, 'name': CreateSnapshot_Task, 'duration_secs': 0.968249} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.164532] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 682.165248] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7958a90-c265-40a6-a881-cb885963691a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.187571] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.224332] env[61728]: DEBUG nova.compute.manager [req-a626d96e-125d-49ef-8b04-8ea5a2d8e45a req-3fc24270-1a0c-4d9e-b6e1-808cc9b24e90 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Received event network-vif-deleted-edc90742-f02d-42ad-85cd-b4242f797c33 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 682.229400] env[61728]: INFO nova.compute.manager [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Took 47.07 seconds to build instance. [ 682.233650] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463879, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100415} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.234259] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 682.235626] env[61728]: DEBUG nova.network.neutron [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.240102] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2869eb9-24a0-491c-bda3-b52f655c797c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.251119] env[61728]: DEBUG oslo_vmware.api [None req-47ff5792-37f3-4772-96dd-d71f681a0954 tempest-ServersAdminNegativeTestJSON-1278040523 tempest-ServersAdminNegativeTestJSON-1278040523-project-admin] Task: {'id': task-463881, 'name': SuspendVM_Task} progress is 91%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.260857] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.378s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.260906] env[61728]: DEBUG nova.compute.manager [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 682.271918] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] b833f691-e76e-4a2e-94a4-7594fadc3ba8/b833f691-e76e-4a2e-94a4-7594fadc3ba8.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 682.275524] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.061s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.278938] env[61728]: INFO nova.compute.claims [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.286192] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-656cc2d1-9cfb-4b82-88a8-e263dfdfd361 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.307169] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 682.307169] env[61728]: value = "task-463883" [ 682.307169] env[61728]: _type = "Task" [ 682.307169] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.316341] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463883, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.375078] env[61728]: DEBUG oslo_vmware.api [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-463882, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.273893} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.375998] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 682.376462] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 682.376540] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 682.376762] env[61728]: INFO nova.compute.manager [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Took 1.70 seconds to destroy the instance on the hypervisor. [ 682.377318] env[61728]: DEBUG oslo.service.loopingcall [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 682.377318] env[61728]: DEBUG nova.compute.manager [-] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 682.377445] env[61728]: DEBUG nova.network.neutron [-] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 682.485993] env[61728]: DEBUG nova.network.neutron [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance_info_cache with network_info: [{"id": "4d70757b-d260-4aa3-8914-a38eb567f7b8", "address": "fa:16:3e:2a:f5:67", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.112", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d70757b-d2", "ovs_interfaceid": "4d70757b-d260-4aa3-8914-a38eb567f7b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.661877] env[61728]: INFO nova.compute.manager [-] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Took 2.26 seconds to deallocate network for instance. [ 682.687603] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 682.687959] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-400057d0-dcd4-4c9f-80b0-d63a74337dde {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.696907] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 682.696907] env[61728]: value = "task-463885" [ 682.696907] env[61728]: _type = "Task" [ 682.696907] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.706158] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463885, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.729800] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a2504bef-cce7-4a13-900c-dddfc7e77595 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Lock "ad304c11-4ff8-4910-92a5-24216d727c4b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.080s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.748987] env[61728]: DEBUG oslo_vmware.api [None req-47ff5792-37f3-4772-96dd-d71f681a0954 tempest-ServersAdminNegativeTestJSON-1278040523 tempest-ServersAdminNegativeTestJSON-1278040523-project-admin] Task: {'id': task-463881, 'name': SuspendVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.777298] env[61728]: DEBUG nova.compute.utils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 682.779955] env[61728]: DEBUG nova.compute.manager [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 682.783814] env[61728]: DEBUG nova.network.neutron [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 682.820164] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463883, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.868091] env[61728]: DEBUG nova.policy [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aac08b5321344d5ab384a3714e6e0bc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '383f38592de14b4ba93fa9a6eeec40aa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 682.990250] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Releasing lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.990610] env[61728]: DEBUG nova.compute.manager [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Instance network_info: |[{"id": "4d70757b-d260-4aa3-8914-a38eb567f7b8", "address": "fa:16:3e:2a:f5:67", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.112", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d70757b-d2", "ovs_interfaceid": "4d70757b-d260-4aa3-8914-a38eb567f7b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 682.991078] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:f5:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '457c42cd-4ddb-4374-923e-d419b7f6eaff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4d70757b-d260-4aa3-8914-a38eb567f7b8', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 682.999781] env[61728]: DEBUG oslo.service.loopingcall [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.000030] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 683.000505] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-001a8df0-3877-4d55-a04d-10ab0fe83ebd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.022910] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 683.022910] env[61728]: value = "task-463886" [ 683.022910] env[61728]: _type = "Task" [ 683.022910] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.031446] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463886, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.044439] env[61728]: DEBUG oslo_concurrency.lockutils [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "dc8a587f-3dc8-43b3-b986-2096e3c4e1f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.044439] env[61728]: DEBUG oslo_concurrency.lockutils [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "dc8a587f-3dc8-43b3-b986-2096e3c4e1f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.169814] env[61728]: DEBUG oslo_concurrency.lockutils [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.213704] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463885, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.233681] env[61728]: DEBUG nova.compute.manager [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 683.247506] env[61728]: DEBUG oslo_vmware.api [None req-47ff5792-37f3-4772-96dd-d71f681a0954 tempest-ServersAdminNegativeTestJSON-1278040523 tempest-ServersAdminNegativeTestJSON-1278040523-project-admin] Task: {'id': task-463881, 'name': SuspendVM_Task, 'duration_secs': 1.045279} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.248481] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-47ff5792-37f3-4772-96dd-d71f681a0954 tempest-ServersAdminNegativeTestJSON-1278040523 tempest-ServersAdminNegativeTestJSON-1278040523-project-admin] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Suspended the VM {{(pid=61728) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 683.249018] env[61728]: DEBUG nova.compute.manager [None req-47ff5792-37f3-4772-96dd-d71f681a0954 tempest-ServersAdminNegativeTestJSON-1278040523 tempest-ServersAdminNegativeTestJSON-1278040523-project-admin] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 683.249975] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909a5c28-5b6d-4fb0-802c-08dac06213ef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.284836] env[61728]: DEBUG nova.compute.manager [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 683.322194] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463883, 'name': ReconfigVM_Task, 'duration_secs': 0.515914} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.322891] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Reconfigured VM instance instance-0000002a to attach disk [datastore1] b833f691-e76e-4a2e-94a4-7594fadc3ba8/b833f691-e76e-4a2e-94a4-7594fadc3ba8.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 683.323363] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9017a16f-8718-4fa8-8dfa-65570acff6a2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.332135] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 683.332135] env[61728]: value = "task-463887" [ 683.332135] env[61728]: _type = "Task" [ 683.332135] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.339095] env[61728]: DEBUG nova.network.neutron [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Successfully created port: ec661d56-ab1e-4593-8990-bf9e698c8880 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 683.345767] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463887, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.534087] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463886, 'name': CreateVM_Task, 'duration_secs': 0.458284} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.534840] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 683.535665] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.535803] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.536216] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 683.540803] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-415e5bac-dc8c-41c3-b50c-e066533f864b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.545850] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 683.545850] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527bca43-75e2-5420-8916-ab73b66793e8" [ 683.545850] env[61728]: _type = "Task" [ 683.545850] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.554708] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527bca43-75e2-5420-8916-ab73b66793e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.712058] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463885, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.760719] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.847156] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463887, 'name': Rename_Task, 'duration_secs': 0.19012} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.850143] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 683.851154] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0945e60b-20c1-4c1c-bf40-94a986d998de {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.859621] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 683.859621] env[61728]: value = "task-463888" [ 683.859621] env[61728]: _type = "Task" [ 683.859621] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.872501] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463888, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.907386] env[61728]: DEBUG oslo_concurrency.lockutils [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "2e83145a-2481-411d-b037-a305a209074e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.907642] env[61728]: DEBUG oslo_concurrency.lockutils [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "2e83145a-2481-411d-b037-a305a209074e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.933912] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897755ce-0d9c-4558-83ab-136492cad17e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.941973] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52e406d-87ea-42f5-8fcf-a516dc1f75b6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.973472] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a43fff7-ead3-408a-9f7c-9d0e78878516 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.981008] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae485ef-9ec2-44f8-92ec-acf51ee55e7e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.995186] env[61728]: DEBUG nova.compute.provider_tree [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.056991] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527bca43-75e2-5420-8916-ab73b66793e8, 'name': SearchDatastore_Task, 'duration_secs': 0.013322} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.057338] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.057574] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 684.057816] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.058994] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.058994] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 684.058994] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7fdc0f91-49ae-4a9c-b239-6223b6b3f6e0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.066586] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 684.066786] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 684.067594] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-542811f6-ef89-46cb-ace3-2a2ebad68e1b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.073269] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 684.073269] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5228f8f1-c714-c29b-6202-c443bccb294f" [ 684.073269] env[61728]: _type = "Task" [ 684.073269] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.081261] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5228f8f1-c714-c29b-6202-c443bccb294f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.102821] env[61728]: DEBUG nova.network.neutron [-] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.209827] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463885, 'name': CloneVM_Task, 'duration_secs': 1.411621} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.210513] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Created linked-clone VM from snapshot [ 684.211787] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2848a87-f483-4b2c-93f6-097c66c1adc1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.222116] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Uploading image f71e7443-ffb9-45db-b5a5-928244822e10 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 684.252822] env[61728]: DEBUG oslo_vmware.rw_handles [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 684.252822] env[61728]: value = "vm-122047" [ 684.252822] env[61728]: _type = "VirtualMachine" [ 684.252822] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 684.253205] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3cc8b39b-409b-4fed-a799-85d699b7d9ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.259875] env[61728]: DEBUG oslo_vmware.rw_handles [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lease: (returnval){ [ 684.259875] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52789674-6baa-5d03-a0b0-0cdb2207cf54" [ 684.259875] env[61728]: _type = "HttpNfcLease" [ 684.259875] env[61728]: } obtained for exporting VM: (result){ [ 684.259875] env[61728]: value = "vm-122047" [ 684.259875] env[61728]: _type = "VirtualMachine" [ 684.259875] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 684.260227] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the lease: (returnval){ [ 684.260227] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52789674-6baa-5d03-a0b0-0cdb2207cf54" [ 684.260227] env[61728]: _type = "HttpNfcLease" [ 684.260227] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 684.266700] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 684.266700] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52789674-6baa-5d03-a0b0-0cdb2207cf54" [ 684.266700] env[61728]: _type = "HttpNfcLease" [ 684.266700] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 684.302275] env[61728]: DEBUG nova.compute.manager [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 684.330316] env[61728]: DEBUG nova.virt.hardware [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 684.330578] env[61728]: DEBUG nova.virt.hardware [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 684.330742] env[61728]: DEBUG nova.virt.hardware [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 684.330929] env[61728]: DEBUG nova.virt.hardware [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 684.331095] env[61728]: DEBUG nova.virt.hardware [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 684.331265] env[61728]: DEBUG nova.virt.hardware [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 684.331479] env[61728]: DEBUG nova.virt.hardware [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 684.331688] env[61728]: DEBUG nova.virt.hardware [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 684.331876] env[61728]: DEBUG nova.virt.hardware [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 684.332059] env[61728]: DEBUG nova.virt.hardware [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 684.332247] env[61728]: DEBUG nova.virt.hardware [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.333180] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de21a615-1542-434e-85d3-38731af38a37 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.341232] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8a4fed-6d52-43cd-a812-6b7c1d7e873f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.368761] env[61728]: DEBUG oslo_vmware.api [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463888, 'name': PowerOnVM_Task, 'duration_secs': 0.48999} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.368954] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 684.369184] env[61728]: INFO nova.compute.manager [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Took 8.30 seconds to spawn the instance on the hypervisor. [ 684.369365] env[61728]: DEBUG nova.compute.manager [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 684.370117] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da99852-db4f-4e99-8ff9-5b5739e2667e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.385423] env[61728]: DEBUG nova.compute.manager [req-113f4b24-5a2b-4b01-b4ec-4dce18fc7c7f req-296d1049-0740-44a8-8bbc-e27e6283465f service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Received event network-vif-deleted-b5b01c97-ed6d-41f8-98e3-9489d2e04f12 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 684.385647] env[61728]: DEBUG nova.compute.manager [req-113f4b24-5a2b-4b01-b4ec-4dce18fc7c7f req-296d1049-0740-44a8-8bbc-e27e6283465f service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Received event network-vif-deleted-48764907-e12d-424f-8b5b-20c70e244c6a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 684.386072] env[61728]: DEBUG nova.compute.manager [req-113f4b24-5a2b-4b01-b4ec-4dce18fc7c7f req-296d1049-0740-44a8-8bbc-e27e6283465f service nova] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Received event network-vif-deleted-b24101a4-7e85-4510-b0ce-1ebede43eb5c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 684.497746] env[61728]: DEBUG nova.scheduler.client.report [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.584129] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5228f8f1-c714-c29b-6202-c443bccb294f, 'name': SearchDatastore_Task, 'duration_secs': 0.011422} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.584928] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57bb9848-46ac-47eb-b61c-fef81792fd34 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.590414] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 684.590414] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5260a310-5a44-953d-ad13-00ef8cdef679" [ 684.590414] env[61728]: _type = "Task" [ 684.590414] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.598466] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5260a310-5a44-953d-ad13-00ef8cdef679, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.606025] env[61728]: INFO nova.compute.manager [-] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Took 2.23 seconds to deallocate network for instance. [ 684.744029] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 684.744029] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122045', 'volume_id': '10d75c82-d308-41da-a8bd-8beaea48adf9', 'name': 'volume-10d75c82-d308-41da-a8bd-8beaea48adf9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '507b5aba-aec9-438b-92c0-5efa17959bc7', 'attached_at': '', 'detached_at': '', 'volume_id': '10d75c82-d308-41da-a8bd-8beaea48adf9', 'serial': '10d75c82-d308-41da-a8bd-8beaea48adf9'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 684.744029] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c5cea3-c658-4acb-8d97-2d3cf4cdf831 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.765032] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6af0a11-8746-4da1-ae73-5f6bcc265c50 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.771023] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 684.771023] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52789674-6baa-5d03-a0b0-0cdb2207cf54" [ 684.771023] env[61728]: _type = "HttpNfcLease" [ 684.771023] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 684.784728] env[61728]: DEBUG oslo_vmware.rw_handles [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 684.784728] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52789674-6baa-5d03-a0b0-0cdb2207cf54" [ 684.784728] env[61728]: _type = "HttpNfcLease" [ 684.784728] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 684.793110] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] volume-10d75c82-d308-41da-a8bd-8beaea48adf9/volume-10d75c82-d308-41da-a8bd-8beaea48adf9.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 684.793995] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffeabfe4-1b00-4494-9079-90778b6b7df6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.796843] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbe7112e-9c0a-4b93-a22b-4fcf3cd7e20c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.816202] env[61728]: DEBUG oslo_vmware.rw_handles [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526e56e4-35ae-3a1e-ac53-b28b173940c6/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 684.816417] env[61728]: DEBUG oslo_vmware.rw_handles [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526e56e4-35ae-3a1e-ac53-b28b173940c6/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 684.818768] env[61728]: DEBUG oslo_vmware.api [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 684.818768] env[61728]: value = "task-463890" [ 684.818768] env[61728]: _type = "Task" [ 684.818768] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.886193] env[61728]: DEBUG oslo_vmware.api [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463890, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.890984] env[61728]: INFO nova.compute.manager [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Took 47.96 seconds to build instance. [ 684.942583] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a778300c-1ce8-47d2-a2aa-fe253380ce34 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.004660] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.728s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.006322] env[61728]: DEBUG nova.compute.manager [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 685.009121] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.177s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.009356] env[61728]: DEBUG nova.objects.instance [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lazy-loading 'resources' on Instance uuid 1ef96606-d960-4283-b372-e2bad47f9367 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 685.108992] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5260a310-5a44-953d-ad13-00ef8cdef679, 'name': SearchDatastore_Task, 'duration_secs': 0.00957} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.108992] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.108992] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 1d9114cb-a0af-4874-962f-27237b3c89cc/1d9114cb-a0af-4874-962f-27237b3c89cc.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 685.108992] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5b75b41f-e715-4425-bbee-474dae980aa5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.114320] env[61728]: DEBUG oslo_concurrency.lockutils [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.114661] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 685.114661] env[61728]: value = "task-463891" [ 685.114661] env[61728]: _type = "Task" [ 685.114661] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.125922] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463891, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.227975] env[61728]: DEBUG nova.network.neutron [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Successfully updated port: ec661d56-ab1e-4593-8990-bf9e698c8880 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 685.329884] env[61728]: DEBUG oslo_vmware.api [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463890, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.392247] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ccef00d-14b0-401a-9972-e9b3024e42ea tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "b833f691-e76e-4a2e-94a4-7594fadc3ba8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.801s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.513316] env[61728]: DEBUG nova.compute.utils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 685.514616] env[61728]: DEBUG nova.compute.manager [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 685.514801] env[61728]: DEBUG nova.network.neutron [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 685.596322] env[61728]: DEBUG nova.policy [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea6bda719ad24902bc0bca7e97be10e7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab547a56a1e24b3bb051b7f0c70b4640', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 685.634078] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463891, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.733757] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "refresh_cache-ca8d41dc-5719-4ade-b82a-b1d7f1c8def1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.733842] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquired lock "refresh_cache-ca8d41dc-5719-4ade-b82a-b1d7f1c8def1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.734245] env[61728]: DEBUG nova.network.neutron [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 685.837504] env[61728]: DEBUG oslo_vmware.api [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463890, 'name': ReconfigVM_Task, 'duration_secs': 0.658101} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.838809] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Reconfigured VM instance instance-00000027 to attach disk [datastore1] volume-10d75c82-d308-41da-a8bd-8beaea48adf9/volume-10d75c82-d308-41da-a8bd-8beaea48adf9.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 685.850435] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02ff4c6e-6a89-40ce-af0a-dce2405f1ccf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.866366] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa37e9eb-2228-4fe3-a30b-a07dd2d00b3a tempest-ServersAdminTestJSON-1533333059 tempest-ServersAdminTestJSON-1533333059-project-admin] Acquiring lock "refresh_cache-b833f691-e76e-4a2e-94a4-7594fadc3ba8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.866605] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa37e9eb-2228-4fe3-a30b-a07dd2d00b3a tempest-ServersAdminTestJSON-1533333059 tempest-ServersAdminTestJSON-1533333059-project-admin] Acquired lock "refresh_cache-b833f691-e76e-4a2e-94a4-7594fadc3ba8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.867119] env[61728]: DEBUG nova.network.neutron [None req-aa37e9eb-2228-4fe3-a30b-a07dd2d00b3a tempest-ServersAdminTestJSON-1533333059 tempest-ServersAdminTestJSON-1533333059-project-admin] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 685.876087] env[61728]: DEBUG oslo_vmware.api [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 685.876087] env[61728]: value = "task-463892" [ 685.876087] env[61728]: _type = "Task" [ 685.876087] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.891048] env[61728]: DEBUG oslo_vmware.api [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463892, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.895765] env[61728]: DEBUG nova.compute.manager [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 686.027246] env[61728]: DEBUG nova.compute.utils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 686.031165] env[61728]: DEBUG nova.network.neutron [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Successfully created port: 613976dd-b5d2-4919-9887-b3d42dc84c87 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.134729] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463891, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.233930] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f825153-4f6f-4684-919b-a8b2d088de25 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.244615] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc5c1012-4186-4a3d-934d-396eb0715cad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.279748] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19c3e40-b861-478e-9fe5-04beda104736 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.283636] env[61728]: DEBUG nova.network.neutron [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.292591] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55eb326d-21f8-4e80-84cc-a3169b8d38c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.308953] env[61728]: DEBUG nova.compute.provider_tree [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.387014] env[61728]: DEBUG oslo_vmware.api [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463892, 'name': ReconfigVM_Task, 'duration_secs': 0.168931} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.387411] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122045', 'volume_id': '10d75c82-d308-41da-a8bd-8beaea48adf9', 'name': 'volume-10d75c82-d308-41da-a8bd-8beaea48adf9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '507b5aba-aec9-438b-92c0-5efa17959bc7', 'attached_at': '', 'detached_at': '', 'volume_id': '10d75c82-d308-41da-a8bd-8beaea48adf9', 'serial': '10d75c82-d308-41da-a8bd-8beaea48adf9'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 686.417149] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.531337] env[61728]: DEBUG nova.compute.manager [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 686.541984] env[61728]: DEBUG nova.network.neutron [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Updating instance_info_cache with network_info: [{"id": "ec661d56-ab1e-4593-8990-bf9e698c8880", "address": "fa:16:3e:b0:10:d0", "network": {"id": "17a8ea01-7320-4089-9138-a6fd17ac4054", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-599506352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "383f38592de14b4ba93fa9a6eeec40aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec661d56-ab", "ovs_interfaceid": "ec661d56-ab1e-4593-8990-bf9e698c8880", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.629992] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463891, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.652493] env[61728]: DEBUG nova.network.neutron [None req-aa37e9eb-2228-4fe3-a30b-a07dd2d00b3a tempest-ServersAdminTestJSON-1533333059 tempest-ServersAdminTestJSON-1533333059-project-admin] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Updating instance_info_cache with network_info: [{"id": "a808f8cb-f136-434e-8bf7-5ee0b957e156", "address": "fa:16:3e:63:52:85", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa808f8cb-f1", "ovs_interfaceid": "a808f8cb-f136-434e-8bf7-5ee0b957e156", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.818285] env[61728]: DEBUG nova.scheduler.client.report [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.997184] env[61728]: DEBUG nova.compute.manager [req-5e1999a0-d314-4448-b97e-a6b16ea759bb req-79f39a8b-865e-41ee-936c-546865be0336 service nova] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Received event network-vif-plugged-ec661d56-ab1e-4593-8990-bf9e698c8880 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 686.997788] env[61728]: DEBUG oslo_concurrency.lockutils [req-5e1999a0-d314-4448-b97e-a6b16ea759bb req-79f39a8b-865e-41ee-936c-546865be0336 service nova] Acquiring lock "ca8d41dc-5719-4ade-b82a-b1d7f1c8def1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.997788] env[61728]: DEBUG oslo_concurrency.lockutils [req-5e1999a0-d314-4448-b97e-a6b16ea759bb req-79f39a8b-865e-41ee-936c-546865be0336 service nova] Lock "ca8d41dc-5719-4ade-b82a-b1d7f1c8def1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.999853] env[61728]: DEBUG oslo_concurrency.lockutils [req-5e1999a0-d314-4448-b97e-a6b16ea759bb req-79f39a8b-865e-41ee-936c-546865be0336 service nova] Lock "ca8d41dc-5719-4ade-b82a-b1d7f1c8def1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.999853] env[61728]: DEBUG nova.compute.manager [req-5e1999a0-d314-4448-b97e-a6b16ea759bb req-79f39a8b-865e-41ee-936c-546865be0336 service nova] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] No waiting events found dispatching network-vif-plugged-ec661d56-ab1e-4593-8990-bf9e698c8880 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 686.999853] env[61728]: WARNING nova.compute.manager [req-5e1999a0-d314-4448-b97e-a6b16ea759bb req-79f39a8b-865e-41ee-936c-546865be0336 service nova] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Received unexpected event network-vif-plugged-ec661d56-ab1e-4593-8990-bf9e698c8880 for instance with vm_state building and task_state spawning. [ 686.999853] env[61728]: DEBUG nova.compute.manager [req-5e1999a0-d314-4448-b97e-a6b16ea759bb req-79f39a8b-865e-41ee-936c-546865be0336 service nova] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Received event network-changed-ec661d56-ab1e-4593-8990-bf9e698c8880 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 686.999853] env[61728]: DEBUG nova.compute.manager [req-5e1999a0-d314-4448-b97e-a6b16ea759bb req-79f39a8b-865e-41ee-936c-546865be0336 service nova] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Refreshing instance network info cache due to event network-changed-ec661d56-ab1e-4593-8990-bf9e698c8880. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 686.999853] env[61728]: DEBUG oslo_concurrency.lockutils [req-5e1999a0-d314-4448-b97e-a6b16ea759bb req-79f39a8b-865e-41ee-936c-546865be0336 service nova] Acquiring lock "refresh_cache-ca8d41dc-5719-4ade-b82a-b1d7f1c8def1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.046562] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Releasing lock "refresh_cache-ca8d41dc-5719-4ade-b82a-b1d7f1c8def1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.047084] env[61728]: DEBUG nova.compute.manager [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Instance network_info: |[{"id": "ec661d56-ab1e-4593-8990-bf9e698c8880", "address": "fa:16:3e:b0:10:d0", "network": {"id": "17a8ea01-7320-4089-9138-a6fd17ac4054", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-599506352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "383f38592de14b4ba93fa9a6eeec40aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec661d56-ab", "ovs_interfaceid": "ec661d56-ab1e-4593-8990-bf9e698c8880", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 687.047266] env[61728]: DEBUG oslo_concurrency.lockutils [req-5e1999a0-d314-4448-b97e-a6b16ea759bb req-79f39a8b-865e-41ee-936c-546865be0336 service nova] Acquired lock "refresh_cache-ca8d41dc-5719-4ade-b82a-b1d7f1c8def1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.047403] env[61728]: DEBUG nova.network.neutron [req-5e1999a0-d314-4448-b97e-a6b16ea759bb req-79f39a8b-865e-41ee-936c-546865be0336 service nova] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Refreshing network info cache for port ec661d56-ab1e-4593-8990-bf9e698c8880 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 687.049317] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:10:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec661d56-ab1e-4593-8990-bf9e698c8880', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 687.058857] env[61728]: DEBUG oslo.service.loopingcall [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 687.059942] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 687.060222] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-51745915-d447-4353-ae93-80f1427d4f00 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.081198] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 687.081198] env[61728]: value = "task-463893" [ 687.081198] env[61728]: _type = "Task" [ 687.081198] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.094389] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463893, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.133908] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463891, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.5765} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.134227] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 1d9114cb-a0af-4874-962f-27237b3c89cc/1d9114cb-a0af-4874-962f-27237b3c89cc.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 687.135063] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 687.135063] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6fd84257-6eb8-40a3-bf96-df77f92d5ebc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.145072] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 687.145072] env[61728]: value = "task-463894" [ 687.145072] env[61728]: _type = "Task" [ 687.145072] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.157239] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa37e9eb-2228-4fe3-a30b-a07dd2d00b3a tempest-ServersAdminTestJSON-1533333059 tempest-ServersAdminTestJSON-1533333059-project-admin] Releasing lock "refresh_cache-b833f691-e76e-4a2e-94a4-7594fadc3ba8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.157623] env[61728]: DEBUG nova.compute.manager [None req-aa37e9eb-2228-4fe3-a30b-a07dd2d00b3a tempest-ServersAdminTestJSON-1533333059 tempest-ServersAdminTestJSON-1533333059-project-admin] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Inject network info {{(pid=61728) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 687.157934] env[61728]: DEBUG nova.compute.manager [None req-aa37e9eb-2228-4fe3-a30b-a07dd2d00b3a tempest-ServersAdminTestJSON-1533333059 tempest-ServersAdminTestJSON-1533333059-project-admin] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] network_info to inject: |[{"id": "a808f8cb-f136-434e-8bf7-5ee0b957e156", "address": "fa:16:3e:63:52:85", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa808f8cb-f1", "ovs_interfaceid": "a808f8cb-f136-434e-8bf7-5ee0b957e156", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 687.164370] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa37e9eb-2228-4fe3-a30b-a07dd2d00b3a tempest-ServersAdminTestJSON-1533333059 tempest-ServersAdminTestJSON-1533333059-project-admin] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Reconfiguring VM instance to set the machine id {{(pid=61728) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 687.164708] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463894, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.164984] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7a249e0-4315-4424-9f8d-844e70be4c87 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.182159] env[61728]: DEBUG oslo_vmware.api [None req-aa37e9eb-2228-4fe3-a30b-a07dd2d00b3a tempest-ServersAdminTestJSON-1533333059 tempest-ServersAdminTestJSON-1533333059-project-admin] Waiting for the task: (returnval){ [ 687.182159] env[61728]: value = "task-463895" [ 687.182159] env[61728]: _type = "Task" [ 687.182159] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.193518] env[61728]: DEBUG oslo_vmware.api [None req-aa37e9eb-2228-4fe3-a30b-a07dd2d00b3a tempest-ServersAdminTestJSON-1533333059 tempest-ServersAdminTestJSON-1533333059-project-admin] Task: {'id': task-463895, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.325561] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.316s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.328522] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.473s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.330169] env[61728]: INFO nova.compute.claims [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.361938] env[61728]: INFO nova.scheduler.client.report [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleted allocations for instance 1ef96606-d960-4283-b372-e2bad47f9367 [ 687.442482] env[61728]: DEBUG nova.objects.instance [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lazy-loading 'flavor' on Instance uuid 507b5aba-aec9-438b-92c0-5efa17959bc7 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 687.546493] env[61728]: DEBUG nova.compute.manager [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 687.580955] env[61728]: DEBUG nova.virt.hardware [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:22:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=1,extra_specs={hw_rng:allowed='True'},flavorid='1584216751',id=30,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_1-930809863',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 687.581487] env[61728]: DEBUG nova.virt.hardware [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 687.581781] env[61728]: DEBUG nova.virt.hardware [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.582050] env[61728]: DEBUG nova.virt.hardware [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 687.582357] env[61728]: DEBUG nova.virt.hardware [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.582607] env[61728]: DEBUG nova.virt.hardware [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 687.583134] env[61728]: DEBUG nova.virt.hardware [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 687.584048] env[61728]: DEBUG nova.virt.hardware [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 687.584293] env[61728]: DEBUG nova.virt.hardware [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 687.584543] env[61728]: DEBUG nova.virt.hardware [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 687.584788] env[61728]: DEBUG nova.virt.hardware [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 687.587358] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b8f0cc-ed24-4e93-aafb-214c2ce138f3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.606952] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-840d18d3-f1a1-4d2f-b7a0-accefa371319 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.611175] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463893, 'name': CreateVM_Task, 'duration_secs': 0.410038} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.613622] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 687.614740] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.614976] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.615294] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 687.624769] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2a778e1-976c-45c5-973d-eb43a8b6184e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.631215] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 687.631215] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ce337d-b8f2-c071-92bd-de5d0da02f00" [ 687.631215] env[61728]: _type = "Task" [ 687.631215] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.641886] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ce337d-b8f2-c071-92bd-de5d0da02f00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.654190] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463894, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0767} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.654587] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 687.655422] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31987fd0-016c-4d42-95bb-f07a367f2f86 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.679306] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] 1d9114cb-a0af-4874-962f-27237b3c89cc/1d9114cb-a0af-4874-962f-27237b3c89cc.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 687.679809] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-836e6dff-e80e-4aa0-913f-fc23f47f5ea0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.703610] env[61728]: DEBUG oslo_vmware.api [None req-aa37e9eb-2228-4fe3-a30b-a07dd2d00b3a tempest-ServersAdminTestJSON-1533333059 tempest-ServersAdminTestJSON-1533333059-project-admin] Task: {'id': task-463895, 'name': ReconfigVM_Task, 'duration_secs': 0.182026} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.707875] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa37e9eb-2228-4fe3-a30b-a07dd2d00b3a tempest-ServersAdminTestJSON-1533333059 tempest-ServersAdminTestJSON-1533333059-project-admin] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Reconfigured VM instance to set the machine id {{(pid=61728) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 687.708447] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 687.708447] env[61728]: value = "task-463896" [ 687.708447] env[61728]: _type = "Task" [ 687.708447] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.722018] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463896, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.779228] env[61728]: DEBUG oslo_concurrency.lockutils [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "650d2875-68f7-4131-b9b7-9799ee08f1d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.779558] env[61728]: DEBUG oslo_concurrency.lockutils [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "650d2875-68f7-4131-b9b7-9799ee08f1d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.779754] env[61728]: DEBUG oslo_concurrency.lockutils [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "650d2875-68f7-4131-b9b7-9799ee08f1d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.779964] env[61728]: DEBUG oslo_concurrency.lockutils [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "650d2875-68f7-4131-b9b7-9799ee08f1d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.780163] env[61728]: DEBUG oslo_concurrency.lockutils [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "650d2875-68f7-4131-b9b7-9799ee08f1d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.782541] env[61728]: INFO nova.compute.manager [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Terminating instance [ 687.784612] env[61728]: DEBUG nova.compute.manager [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 687.784904] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 687.786253] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ce1688-0c12-46ff-ba9a-bfb64610a71c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.795084] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 687.795490] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-08fb2683-8981-40f4-8104-2765091cf6d1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.861238] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 687.861783] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 687.861990] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Deleting the datastore file [datastore1] 650d2875-68f7-4131-b9b7-9799ee08f1d6 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 687.862281] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17898c9d-f2f2-4aaf-ab53-db0af281b2d0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.873091] env[61728]: DEBUG oslo_vmware.api [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 687.873091] env[61728]: value = "task-463898" [ 687.873091] env[61728]: _type = "Task" [ 687.873091] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.873639] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d36e90f8-e370-4098-b3f9-11dfde17398c tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "1ef96606-d960-4283-b372-e2bad47f9367" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.287s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.885340] env[61728]: DEBUG oslo_vmware.api [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463898, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.947804] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8b9a7d14-9176-4a68-bdb1-35c51d071ced tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "507b5aba-aec9-438b-92c0-5efa17959bc7" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.831s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.958879] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "507b5aba-aec9-438b-92c0-5efa17959bc7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.958879] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "507b5aba-aec9-438b-92c0-5efa17959bc7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.958879] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "507b5aba-aec9-438b-92c0-5efa17959bc7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.958879] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "507b5aba-aec9-438b-92c0-5efa17959bc7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.959151] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "507b5aba-aec9-438b-92c0-5efa17959bc7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.961832] env[61728]: INFO nova.compute.manager [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Terminating instance [ 687.965138] env[61728]: DEBUG nova.compute.manager [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 687.965456] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 687.966738] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11de60b5-1fc8-479f-a118-b7d581aa744a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.968899] env[61728]: DEBUG nova.network.neutron [req-5e1999a0-d314-4448-b97e-a6b16ea759bb req-79f39a8b-865e-41ee-936c-546865be0336 service nova] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Updated VIF entry in instance network info cache for port ec661d56-ab1e-4593-8990-bf9e698c8880. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 687.969249] env[61728]: DEBUG nova.network.neutron [req-5e1999a0-d314-4448-b97e-a6b16ea759bb req-79f39a8b-865e-41ee-936c-546865be0336 service nova] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Updating instance_info_cache with network_info: [{"id": "ec661d56-ab1e-4593-8990-bf9e698c8880", "address": "fa:16:3e:b0:10:d0", "network": {"id": "17a8ea01-7320-4089-9138-a6fd17ac4054", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-599506352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "383f38592de14b4ba93fa9a6eeec40aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec661d56-ab", "ovs_interfaceid": "ec661d56-ab1e-4593-8990-bf9e698c8880", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.976718] env[61728]: DEBUG oslo_vmware.api [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 687.976718] env[61728]: value = "task-463899" [ 687.976718] env[61728]: _type = "Task" [ 687.976718] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.980712] env[61728]: DEBUG nova.network.neutron [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Successfully updated port: 613976dd-b5d2-4919-9887-b3d42dc84c87 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 687.988285] env[61728]: DEBUG oslo_vmware.api [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463899, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.142603] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ce337d-b8f2-c071-92bd-de5d0da02f00, 'name': SearchDatastore_Task, 'duration_secs': 0.012096} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.142926] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.143287] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 688.143411] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.143560] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.143749] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 688.144037] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-395fec97-6118-4711-9150-513b33d5c40e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.166150] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 688.166511] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 688.167214] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f5324cb-7ebe-40d0-b1ba-738c8c192bdb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.173272] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 688.173272] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52431515-5119-bd6c-7f1c-b28abcf05d7e" [ 688.173272] env[61728]: _type = "Task" [ 688.173272] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.183825] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52431515-5119-bd6c-7f1c-b28abcf05d7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.217488] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.317205] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Acquiring lock "ad304c11-4ff8-4910-92a5-24216d727c4b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.317205] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Lock "ad304c11-4ff8-4910-92a5-24216d727c4b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.317578] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Acquiring lock "ad304c11-4ff8-4910-92a5-24216d727c4b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.318404] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Lock "ad304c11-4ff8-4910-92a5-24216d727c4b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.318404] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Lock "ad304c11-4ff8-4910-92a5-24216d727c4b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.320702] env[61728]: INFO nova.compute.manager [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Terminating instance [ 688.322867] env[61728]: DEBUG nova.compute.manager [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 688.323171] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 688.324217] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e6d933-8f0a-4074-adb1-c88bdb2f116b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.333590] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 688.333981] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dbad2109-7f9c-4dde-8eb7-97bdc3f095af {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.343443] env[61728]: DEBUG oslo_vmware.api [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Waiting for the task: (returnval){ [ 688.343443] env[61728]: value = "task-463900" [ 688.343443] env[61728]: _type = "Task" [ 688.343443] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.353743] env[61728]: DEBUG oslo_vmware.api [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463900, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.388032] env[61728]: DEBUG oslo_vmware.api [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-463898, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.316828} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.391266] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 688.391962] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 688.392228] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 688.392457] env[61728]: INFO nova.compute.manager [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Took 0.61 seconds to destroy the instance on the hypervisor. [ 688.392871] env[61728]: DEBUG oslo.service.loopingcall [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 688.393241] env[61728]: DEBUG nova.compute.manager [-] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 688.393382] env[61728]: DEBUG nova.network.neutron [-] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 688.478021] env[61728]: DEBUG oslo_concurrency.lockutils [req-5e1999a0-d314-4448-b97e-a6b16ea759bb req-79f39a8b-865e-41ee-936c-546865be0336 service nova] Releasing lock "refresh_cache-ca8d41dc-5719-4ade-b82a-b1d7f1c8def1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.485411] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "refresh_cache-b563e521-0e0e-45ea-bd5a-056285ea35b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.485411] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquired lock "refresh_cache-b563e521-0e0e-45ea-bd5a-056285ea35b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.486178] env[61728]: DEBUG nova.network.neutron [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 688.491474] env[61728]: DEBUG oslo_vmware.api [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463899, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.691367] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52431515-5119-bd6c-7f1c-b28abcf05d7e, 'name': SearchDatastore_Task, 'duration_secs': 0.018192} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.692318] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28e350a1-78f8-42ca-9564-3ee4846cfc8c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.698747] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 688.698747] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e5cc66-7ef1-9677-fb50-b90f00ef954a" [ 688.698747] env[61728]: _type = "Task" [ 688.698747] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.720546] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e5cc66-7ef1-9677-fb50-b90f00ef954a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.730029] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.858640] env[61728]: DEBUG oslo_vmware.api [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463900, 'name': PowerOffVM_Task, 'duration_secs': 0.222118} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.858897] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 688.859084] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 688.859436] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47d3944b-d83a-4d0b-abd2-cbca032d1427 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.927232] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 688.927638] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 688.927929] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Deleting the datastore file [datastore1] ad304c11-4ff8-4910-92a5-24216d727c4b {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 688.928344] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-01849923-382b-4f18-bf4f-5b00e2e204db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.939245] env[61728]: DEBUG oslo_vmware.api [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Waiting for the task: (returnval){ [ 688.939245] env[61728]: value = "task-463902" [ 688.939245] env[61728]: _type = "Task" [ 688.939245] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.948968] env[61728]: DEBUG oslo_vmware.api [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463902, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.975237] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b16b1c0-4b7a-4988-8389-b89b817bbdd9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.988554] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19495f8e-3db0-48e1-8d75-15c749f02022 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.992509] env[61728]: DEBUG oslo_vmware.api [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463899, 'name': PowerOffVM_Task, 'duration_secs': 0.675982} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.995397] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 688.995829] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 688.996314] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122045', 'volume_id': '10d75c82-d308-41da-a8bd-8beaea48adf9', 'name': 'volume-10d75c82-d308-41da-a8bd-8beaea48adf9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '507b5aba-aec9-438b-92c0-5efa17959bc7', 'attached_at': '', 'detached_at': '', 'volume_id': '10d75c82-d308-41da-a8bd-8beaea48adf9', 'serial': '10d75c82-d308-41da-a8bd-8beaea48adf9'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 688.998318] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b5ba89-9b9b-49aa-ba5f-8e46a1c5b43b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.032417] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cefe90c-7c02-495b-97d6-ea07759ca2e7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.053525] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f627121-abf1-488d-b4c3-e5123573e384 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.059686] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e6f0c5-c5f1-4026-a513-7be3ba16ddfd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.065977] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5773282f-3d78-4e55-8ac1-ad01ca03c92c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.076553] env[61728]: DEBUG nova.network.neutron [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.082045] env[61728]: DEBUG nova.compute.provider_tree [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.103784] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a8ed8fc-d274-46b3-a874-5d73d62bdaeb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.122222] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] The volume has not been displaced from its original location: [datastore1] volume-10d75c82-d308-41da-a8bd-8beaea48adf9/volume-10d75c82-d308-41da-a8bd-8beaea48adf9.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 689.127565] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Reconfiguring VM instance instance-00000027 to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 689.128269] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-894697fc-88a5-4455-a34e-b8913ed1ac40 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.150222] env[61728]: DEBUG oslo_vmware.api [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 689.150222] env[61728]: value = "task-463903" [ 689.150222] env[61728]: _type = "Task" [ 689.150222] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.162942] env[61728]: DEBUG oslo_vmware.api [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463903, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.211488] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e5cc66-7ef1-9677-fb50-b90f00ef954a, 'name': SearchDatastore_Task, 'duration_secs': 0.017289} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.211854] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.212188] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] ca8d41dc-5719-4ade-b82a-b1d7f1c8def1/ca8d41dc-5719-4ade-b82a-b1d7f1c8def1.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 689.213140] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bd1a96bd-8029-4eec-a181-157c568ebde6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.223879] env[61728]: DEBUG nova.network.neutron [-] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.225292] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463896, 'name': ReconfigVM_Task, 'duration_secs': 1.115918} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.226992] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Reconfigured VM instance instance-0000002b to attach disk [datastore1] 1d9114cb-a0af-4874-962f-27237b3c89cc/1d9114cb-a0af-4874-962f-27237b3c89cc.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 689.227754] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 689.227754] env[61728]: value = "task-463904" [ 689.227754] env[61728]: _type = "Task" [ 689.227754] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.228584] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2a1342d9-987e-48bb-8a50-0e181587ceaa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.240833] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463904, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.243209] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 689.243209] env[61728]: value = "task-463905" [ 689.243209] env[61728]: _type = "Task" [ 689.243209] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.257482] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463905, 'name': Rename_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.260493] env[61728]: DEBUG nova.compute.manager [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Received event network-vif-plugged-613976dd-b5d2-4919-9887-b3d42dc84c87 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 689.260715] env[61728]: DEBUG oslo_concurrency.lockutils [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] Acquiring lock "b563e521-0e0e-45ea-bd5a-056285ea35b9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.260976] env[61728]: DEBUG oslo_concurrency.lockutils [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] Lock "b563e521-0e0e-45ea-bd5a-056285ea35b9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.261138] env[61728]: DEBUG oslo_concurrency.lockutils [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] Lock "b563e521-0e0e-45ea-bd5a-056285ea35b9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.261308] env[61728]: DEBUG nova.compute.manager [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] No waiting events found dispatching network-vif-plugged-613976dd-b5d2-4919-9887-b3d42dc84c87 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 689.261470] env[61728]: WARNING nova.compute.manager [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Received unexpected event network-vif-plugged-613976dd-b5d2-4919-9887-b3d42dc84c87 for instance with vm_state building and task_state spawning. [ 689.261948] env[61728]: DEBUG nova.compute.manager [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Received event network-changed-613976dd-b5d2-4919-9887-b3d42dc84c87 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 689.261948] env[61728]: DEBUG nova.compute.manager [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Refreshing instance network info cache due to event network-changed-613976dd-b5d2-4919-9887-b3d42dc84c87. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 689.262117] env[61728]: DEBUG oslo_concurrency.lockutils [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] Acquiring lock "refresh_cache-b563e521-0e0e-45ea-bd5a-056285ea35b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.309583] env[61728]: DEBUG nova.network.neutron [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Updating instance_info_cache with network_info: [{"id": "613976dd-b5d2-4919-9887-b3d42dc84c87", "address": "fa:16:3e:9b:9b:90", "network": {"id": "113d3399-4dbf-44c8-9fef-d29cbb233c49", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471418633-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab547a56a1e24b3bb051b7f0c70b4640", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap613976dd-b5", "ovs_interfaceid": "613976dd-b5d2-4919-9887-b3d42dc84c87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.448037] env[61728]: DEBUG oslo_vmware.api [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Task: {'id': task-463902, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.435035} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.448404] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 689.451171] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 689.451171] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 689.451171] env[61728]: INFO nova.compute.manager [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 689.451171] env[61728]: DEBUG oslo.service.loopingcall [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.451171] env[61728]: DEBUG nova.compute.manager [-] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.451171] env[61728]: DEBUG nova.network.neutron [-] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 689.588320] env[61728]: DEBUG nova.scheduler.client.report [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 689.664415] env[61728]: DEBUG oslo_vmware.api [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463903, 'name': ReconfigVM_Task, 'duration_secs': 0.328157} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.664415] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Reconfigured VM instance instance-00000027 to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 689.669846] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0d578de-d4dd-4aab-8255-53857f863f86 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.694421] env[61728]: DEBUG oslo_vmware.api [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 689.694421] env[61728]: value = "task-463906" [ 689.694421] env[61728]: _type = "Task" [ 689.694421] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.707767] env[61728]: DEBUG oslo_vmware.api [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463906, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.726936] env[61728]: INFO nova.compute.manager [-] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Took 1.33 seconds to deallocate network for instance. [ 689.777756] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463904, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.778457] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463905, 'name': Rename_Task, 'duration_secs': 0.200658} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.779115] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 689.780225] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74a18c9f-3db9-4fb0-bf96-ef3ceae99c52 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.789732] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 689.789732] env[61728]: value = "task-463907" [ 689.789732] env[61728]: _type = "Task" [ 689.789732] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.806761] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463907, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.812398] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Releasing lock "refresh_cache-b563e521-0e0e-45ea-bd5a-056285ea35b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.813094] env[61728]: DEBUG nova.compute.manager [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Instance network_info: |[{"id": "613976dd-b5d2-4919-9887-b3d42dc84c87", "address": "fa:16:3e:9b:9b:90", "network": {"id": "113d3399-4dbf-44c8-9fef-d29cbb233c49", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471418633-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab547a56a1e24b3bb051b7f0c70b4640", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap613976dd-b5", "ovs_interfaceid": "613976dd-b5d2-4919-9887-b3d42dc84c87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 689.813647] env[61728]: DEBUG oslo_concurrency.lockutils [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] Acquired lock "refresh_cache-b563e521-0e0e-45ea-bd5a-056285ea35b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.814150] env[61728]: DEBUG nova.network.neutron [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Refreshing network info cache for port 613976dd-b5d2-4919-9887-b3d42dc84c87 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 689.816324] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9b:9b:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '459b8c74-0aa6-42b6-996a-42b1c5d7e5c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '613976dd-b5d2-4919-9887-b3d42dc84c87', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 689.831696] env[61728]: DEBUG oslo.service.loopingcall [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.833692] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 689.834242] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82609563-30a5-418b-9a37-e7d92e5752e2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.873328] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 689.873328] env[61728]: value = "task-463908" [ 689.873328] env[61728]: _type = "Task" [ 689.873328] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.889262] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463908, 'name': CreateVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.097026] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.766s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.097026] env[61728]: DEBUG nova.compute.manager [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 690.102213] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.130s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.102616] env[61728]: DEBUG nova.objects.instance [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lazy-loading 'resources' on Instance uuid f947d808-ebea-4593-b8e6-d3851c5216dc {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 690.205588] env[61728]: DEBUG oslo_vmware.api [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463906, 'name': ReconfigVM_Task, 'duration_secs': 0.238693} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.206030] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122045', 'volume_id': '10d75c82-d308-41da-a8bd-8beaea48adf9', 'name': 'volume-10d75c82-d308-41da-a8bd-8beaea48adf9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '507b5aba-aec9-438b-92c0-5efa17959bc7', 'attached_at': '', 'detached_at': '', 'volume_id': '10d75c82-d308-41da-a8bd-8beaea48adf9', 'serial': '10d75c82-d308-41da-a8bd-8beaea48adf9'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 690.206362] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 690.207226] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4794b5e-0d2d-4254-935e-ad2037509be0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.214399] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 690.214663] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23ff931a-ccb0-460e-bdd7-7ce32567974c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.221807] env[61728]: DEBUG nova.network.neutron [-] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.247903] env[61728]: DEBUG oslo_concurrency.lockutils [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.252305] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463904, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.910394} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.253223] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] ca8d41dc-5719-4ade-b82a-b1d7f1c8def1/ca8d41dc-5719-4ade-b82a-b1d7f1c8def1.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 690.253223] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 690.253488] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-605a70db-6a52-4a66-bbe8-92d0a1e9ae93 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.262856] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 690.262856] env[61728]: value = "task-463910" [ 690.262856] env[61728]: _type = "Task" [ 690.262856] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.269814] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463910, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.301411] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463907, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.383521] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463908, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.432902] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 690.433016] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 690.433236] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleting the datastore file [datastore1] 507b5aba-aec9-438b-92c0-5efa17959bc7 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 690.433528] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-10e2ad47-fdec-4de0-98c1-dd878dd64963 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.441513] env[61728]: DEBUG oslo_vmware.api [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 690.441513] env[61728]: value = "task-463911" [ 690.441513] env[61728]: _type = "Task" [ 690.441513] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.452767] env[61728]: DEBUG oslo_vmware.api [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463911, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.559040] env[61728]: DEBUG nova.network.neutron [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Updated VIF entry in instance network info cache for port 613976dd-b5d2-4919-9887-b3d42dc84c87. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 690.559474] env[61728]: DEBUG nova.network.neutron [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Updating instance_info_cache with network_info: [{"id": "613976dd-b5d2-4919-9887-b3d42dc84c87", "address": "fa:16:3e:9b:9b:90", "network": {"id": "113d3399-4dbf-44c8-9fef-d29cbb233c49", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471418633-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab547a56a1e24b3bb051b7f0c70b4640", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap613976dd-b5", "ovs_interfaceid": "613976dd-b5d2-4919-9887-b3d42dc84c87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.607276] env[61728]: DEBUG nova.compute.utils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.615121] env[61728]: DEBUG nova.compute.manager [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 690.615121] env[61728]: DEBUG nova.network.neutron [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 690.665873] env[61728]: DEBUG nova.policy [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd35296efd6b947fba051391fd697764f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '01839050e56e4dbfa6e22471a3c60998', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 690.726024] env[61728]: INFO nova.compute.manager [-] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Took 1.28 seconds to deallocate network for instance. [ 690.774603] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463910, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071323} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.779347] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 690.781599] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec18328-9dbe-44f3-954d-b6a1610932ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.807327] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] ca8d41dc-5719-4ade-b82a-b1d7f1c8def1/ca8d41dc-5719-4ade-b82a-b1d7f1c8def1.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 690.814092] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef83169e-dfc6-4804-9bea-12425936c912 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.836670] env[61728]: DEBUG oslo_vmware.api [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-463907, 'name': PowerOnVM_Task, 'duration_secs': 0.789887} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.838443] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 690.838720] env[61728]: INFO nova.compute.manager [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Took 11.96 seconds to spawn the instance on the hypervisor. [ 690.838955] env[61728]: DEBUG nova.compute.manager [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 690.839376] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 690.839376] env[61728]: value = "task-463912" [ 690.839376] env[61728]: _type = "Task" [ 690.839376] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.840398] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2d872d-8c60-41cd-8820-723c9d41946b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.860894] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463912, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.885137] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463908, 'name': CreateVM_Task, 'duration_secs': 0.695917} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.888059] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 690.889306] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.889459] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.889811] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 690.890209] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-552d193a-6b47-42da-8d9a-4ce73d4d85a5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.895068] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 690.895068] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5202546e-7a2f-ab32-e094-fa5c61ea3f87" [ 690.895068] env[61728]: _type = "Task" [ 690.895068] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.904333] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5202546e-7a2f-ab32-e094-fa5c61ea3f87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.957575] env[61728]: DEBUG oslo_vmware.api [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-463911, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.229312} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.958123] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 690.958290] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 690.958474] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 690.958659] env[61728]: INFO nova.compute.manager [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Took 2.99 seconds to destroy the instance on the hypervisor. [ 690.958925] env[61728]: DEBUG oslo.service.loopingcall [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 690.959143] env[61728]: DEBUG nova.compute.manager [-] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 690.959240] env[61728]: DEBUG nova.network.neutron [-] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 691.064174] env[61728]: DEBUG oslo_concurrency.lockutils [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] Releasing lock "refresh_cache-b563e521-0e0e-45ea-bd5a-056285ea35b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.065126] env[61728]: DEBUG nova.compute.manager [req-a44dea26-9592-4cb7-8936-f169f81329b6 req-9d231051-6b88-4f39-9d48-1b5fa4160afc service nova] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Received event network-vif-deleted-fdd5e1d2-60d3-4304-a31b-e6b680c01c63 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 691.114511] env[61728]: DEBUG nova.compute.manager [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 691.238604] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.259959] env[61728]: DEBUG nova.network.neutron [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Successfully created port: 124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 691.308095] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3efb5a72-f1c6-45d5-8e32-30c608f78931 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.322029] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-225b2041-33f2-48cb-9e3d-f0f4dbc02fe0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.360626] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ddd07a3-989d-4964-8784-92af28a119f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.377992] env[61728]: INFO nova.compute.manager [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Took 44.56 seconds to build instance. [ 691.380165] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ba108f-7e22-4ecd-91cc-6f1c32ea3ed1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.384319] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463912, 'name': ReconfigVM_Task, 'duration_secs': 0.533383} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.384867] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Reconfigured VM instance instance-0000002c to attach disk [datastore1] ca8d41dc-5719-4ade-b82a-b1d7f1c8def1/ca8d41dc-5719-4ade-b82a-b1d7f1c8def1.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 691.386495] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6369c705-3639-4038-94f4-4b8ded67dfbd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.399165] env[61728]: DEBUG nova.compute.provider_tree [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.403355] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 691.403355] env[61728]: value = "task-463913" [ 691.403355] env[61728]: _type = "Task" [ 691.403355] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.411311] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5202546e-7a2f-ab32-e094-fa5c61ea3f87, 'name': SearchDatastore_Task, 'duration_secs': 0.012597} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.412058] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.412298] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 691.412532] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.412713] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.412909] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 691.414063] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8a98c625-628e-49a4-a6c4-020d68f8aeae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.418764] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463913, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.426278] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 691.426482] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 691.427283] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-455cd5c0-b422-4cec-89a0-45bbfb5d1d03 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.433598] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 691.433598] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526db810-44c9-68c4-3350-371561d05c38" [ 691.433598] env[61728]: _type = "Task" [ 691.433598] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.441771] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526db810-44c9-68c4-3350-371561d05c38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.500121] env[61728]: DEBUG nova.compute.manager [req-d4cf09b8-3af0-4a9c-9da5-232f980b6c51 req-e4804b7c-956d-420b-9830-6e5367690ff3 service nova] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Received event network-vif-deleted-2537e303-47b7-4ed7-b8b9-7fd10ffc69d5 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 691.887673] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec9a8dea-a57f-4885-b517-311fd10b0b75 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "1d9114cb-a0af-4874-962f-27237b3c89cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.684s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.905254] env[61728]: DEBUG nova.scheduler.client.report [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 691.919371] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463913, 'name': Rename_Task, 'duration_secs': 0.237832} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.919703] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 691.919972] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26b6ca69-b54f-44f1-8218-59f0058ea46b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.927603] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 691.927603] env[61728]: value = "task-463914" [ 691.927603] env[61728]: _type = "Task" [ 691.927603] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.947326] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463914, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.955881] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526db810-44c9-68c4-3350-371561d05c38, 'name': SearchDatastore_Task, 'duration_secs': 0.013258} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.957265] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2c3ad4a-bea8-4497-9420-584c5444c6da {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.966436] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 691.966436] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5284ec1d-7023-5cfa-20d7-d3e4f0b5276d" [ 691.966436] env[61728]: _type = "Task" [ 691.966436] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.977730] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5284ec1d-7023-5cfa-20d7-d3e4f0b5276d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.078595] env[61728]: DEBUG nova.network.neutron [-] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.130024] env[61728]: DEBUG nova.compute.manager [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 692.164474] env[61728]: DEBUG nova.virt.hardware [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 692.164683] env[61728]: DEBUG nova.virt.hardware [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 692.164894] env[61728]: DEBUG nova.virt.hardware [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.165115] env[61728]: DEBUG nova.virt.hardware [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 692.165520] env[61728]: DEBUG nova.virt.hardware [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.165695] env[61728]: DEBUG nova.virt.hardware [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 692.166174] env[61728]: DEBUG nova.virt.hardware [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 692.166365] env[61728]: DEBUG nova.virt.hardware [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 692.166684] env[61728]: DEBUG nova.virt.hardware [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 692.166822] env[61728]: DEBUG nova.virt.hardware [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 692.166881] env[61728]: DEBUG nova.virt.hardware [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 692.168151] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cbab429-b3fb-4bcf-a7c5-f938b0a55647 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.178183] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a299f5-e9a2-4743-9c52-ebaa2fa8eaca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.341163] env[61728]: INFO nova.compute.manager [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Rebuilding instance [ 692.392984] env[61728]: DEBUG nova.compute.manager [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 692.403101] env[61728]: DEBUG nova.compute.manager [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 692.403101] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca9581e-30d0-415a-b087-be67afa760c9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.415768] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.314s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.419535] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.791s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.419863] env[61728]: DEBUG nova.objects.instance [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lazy-loading 'resources' on Instance uuid 4fcc957e-5175-4589-b4da-d8451d92c815 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 692.442043] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463914, 'name': PowerOnVM_Task} progress is 86%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.458290] env[61728]: INFO nova.scheduler.client.report [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Deleted allocations for instance f947d808-ebea-4593-b8e6-d3851c5216dc [ 692.478613] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5284ec1d-7023-5cfa-20d7-d3e4f0b5276d, 'name': SearchDatastore_Task, 'duration_secs': 0.013958} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.478613] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.478898] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b563e521-0e0e-45ea-bd5a-056285ea35b9/b563e521-0e0e-45ea-bd5a-056285ea35b9.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 692.479508] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2627d2bc-ef09-423f-810c-013fc182d29e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.488430] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 692.488430] env[61728]: value = "task-463915" [ 692.488430] env[61728]: _type = "Task" [ 692.488430] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.499975] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463915, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.582034] env[61728]: INFO nova.compute.manager [-] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Took 1.62 seconds to deallocate network for instance. [ 692.917365] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 692.917714] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f2cefea-694d-4746-ad11-3dd2c83d0ff0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.929902] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.930483] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 692.930483] env[61728]: value = "task-463916" [ 692.930483] env[61728]: _type = "Task" [ 692.930483] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.947797] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463914, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.952344] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463916, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.971715] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0ba7188e-054f-4e21-bf37-0f3d416e0b33 tempest-ImagesOneServerTestJSON-6797865 tempest-ImagesOneServerTestJSON-6797865-project-member] Lock "f947d808-ebea-4593-b8e6-d3851c5216dc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.896s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.005434] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463915, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.142113] env[61728]: DEBUG nova.network.neutron [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Successfully updated port: 124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 693.159769] env[61728]: INFO nova.compute.manager [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Took 0.58 seconds to detach 1 volumes for instance. [ 693.455714] env[61728]: DEBUG oslo_vmware.api [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463914, 'name': PowerOnVM_Task, 'duration_secs': 1.12105} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.455714] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463916, 'name': PowerOffVM_Task, 'duration_secs': 0.301045} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.457870] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 693.458147] env[61728]: INFO nova.compute.manager [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Took 9.16 seconds to spawn the instance on the hypervisor. [ 693.458339] env[61728]: DEBUG nova.compute.manager [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 693.458675] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 693.458896] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 693.459974] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abdc863d-0d03-44a7-817e-57922f3118dc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.463261] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba66f977-97b2-4be5-b03b-99d7092bf3ac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.472057] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6e2a1e-7ca7-4301-b1a2-e33f7431eae2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.482038] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 693.483421] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b9da55da-59ff-406a-854f-e67a7332970f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.490047] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77752e48-ce29-4ccf-abf6-9d922ed3c203 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.544410] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463915, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.563954} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.546228] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a325223-261f-44c7-a93c-197462f2200d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.549343] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b563e521-0e0e-45ea-bd5a-056285ea35b9/b563e521-0e0e-45ea-bd5a-056285ea35b9.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 693.549584] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 693.552276] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ad5c2f2-37f6-401c-82aa-06a2d34ada2e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.555144] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 693.555373] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 693.555540] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleting the datastore file [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 693.557400] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4632799e-9abb-4d76-8835-f58bd8e3893d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.564695] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6087e49e-2f84-4503-b608-0902035a3f58 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.572606] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 693.572606] env[61728]: value = "task-463919" [ 693.572606] env[61728]: _type = "Task" [ 693.572606] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.572819] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 693.572819] env[61728]: value = "task-463918" [ 693.572819] env[61728]: _type = "Task" [ 693.572819] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.589156] env[61728]: DEBUG nova.compute.provider_tree [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.599631] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463919, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.599900] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463918, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.645131] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.645354] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquired lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.645562] env[61728]: DEBUG nova.network.neutron [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 693.681261] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.842352] env[61728]: DEBUG nova.compute.manager [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Received event network-vif-deleted-c170e82d-b764-4616-863c-35d45b72edbc {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 693.843503] env[61728]: DEBUG nova.compute.manager [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Received event network-vif-plugged-124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 693.844663] env[61728]: DEBUG oslo_concurrency.lockutils [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] Acquiring lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.844663] env[61728]: DEBUG oslo_concurrency.lockutils [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.844788] env[61728]: DEBUG oslo_concurrency.lockutils [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.844993] env[61728]: DEBUG nova.compute.manager [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] No waiting events found dispatching network-vif-plugged-124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 693.845229] env[61728]: WARNING nova.compute.manager [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Received unexpected event network-vif-plugged-124d4780-15ec-4f17-a2c4-edf1517ed589 for instance with vm_state building and task_state spawning. [ 693.846228] env[61728]: DEBUG nova.compute.manager [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Received event network-changed-124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 693.847075] env[61728]: DEBUG nova.compute.manager [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Refreshing instance network info cache due to event network-changed-124d4780-15ec-4f17-a2c4-edf1517ed589. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 693.847199] env[61728]: DEBUG oslo_concurrency.lockutils [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] Acquiring lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.968102] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "40ef29df-12d6-4274-846a-4a2c32752b8b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.968102] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "40ef29df-12d6-4274-846a-4a2c32752b8b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.995131] env[61728]: INFO nova.compute.manager [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Took 44.66 seconds to build instance. [ 694.101978] env[61728]: DEBUG nova.scheduler.client.report [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.113161] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463918, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.12372} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.114067] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463919, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.306668} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.114399] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 694.114754] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 694.114912] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 694.115108] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 694.119613] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156cffcc-72eb-4e64-b3fe-0d898326ab9a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.150120] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Reconfiguring VM instance instance-0000002d to attach disk [datastore1] b563e521-0e0e-45ea-bd5a-056285ea35b9/b563e521-0e0e-45ea-bd5a-056285ea35b9.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 694.153970] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df513354-0fe9-41ef-8cf9-be820f55579c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.182144] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 694.182144] env[61728]: value = "task-463920" [ 694.182144] env[61728]: _type = "Task" [ 694.182144] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.191516] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463920, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.204052] env[61728]: DEBUG oslo_vmware.rw_handles [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526e56e4-35ae-3a1e-ac53-b28b173940c6/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 694.204996] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e180f2f-5df9-4e7e-aff2-493a131975e8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.213204] env[61728]: DEBUG oslo_vmware.rw_handles [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526e56e4-35ae-3a1e-ac53-b28b173940c6/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 694.213467] env[61728]: ERROR oslo_vmware.rw_handles [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526e56e4-35ae-3a1e-ac53-b28b173940c6/disk-0.vmdk due to incomplete transfer. [ 694.213753] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ec80b311-927e-449a-a107-7f600461c459 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.222270] env[61728]: DEBUG oslo_vmware.rw_handles [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526e56e4-35ae-3a1e-ac53-b28b173940c6/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 694.222270] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Uploaded image f71e7443-ffb9-45db-b5a5-928244822e10 to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 694.223396] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 694.223785] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bcaa2f77-2c56-425a-8d80-a9b2ae4584c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.230127] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 694.230127] env[61728]: value = "task-463921" [ 694.230127] env[61728]: _type = "Task" [ 694.230127] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.237679] env[61728]: DEBUG nova.network.neutron [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 694.250124] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463921, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.490383] env[61728]: DEBUG nova.network.neutron [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Updating instance_info_cache with network_info: [{"id": "124d4780-15ec-4f17-a2c4-edf1517ed589", "address": "fa:16:3e:6d:47:93", "network": {"id": "6a1908cc-f6aa-4dd7-970e-c8a07e1b16ee", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2047209328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "01839050e56e4dbfa6e22471a3c60998", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap124d4780-15", "ovs_interfaceid": "124d4780-15ec-4f17-a2c4-edf1517ed589", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.498781] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b88582-edb8-41bb-80f8-25a1afde8bb1 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "ca8d41dc-5719-4ade-b82a-b1d7f1c8def1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.156s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.615351] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.196s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.618176] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.190s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.619304] env[61728]: DEBUG nova.objects.instance [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Lazy-loading 'resources' on Instance uuid 69d7af16-7708-4df6-acca-093b6c1df1ba {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 694.650131] env[61728]: INFO nova.scheduler.client.report [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Deleted allocations for instance 4fcc957e-5175-4589-b4da-d8451d92c815 [ 694.692974] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463920, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.743631] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463921, 'name': Destroy_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.993404] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Releasing lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.993813] env[61728]: DEBUG nova.compute.manager [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Instance network_info: |[{"id": "124d4780-15ec-4f17-a2c4-edf1517ed589", "address": "fa:16:3e:6d:47:93", "network": {"id": "6a1908cc-f6aa-4dd7-970e-c8a07e1b16ee", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2047209328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "01839050e56e4dbfa6e22471a3c60998", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap124d4780-15", "ovs_interfaceid": "124d4780-15ec-4f17-a2c4-edf1517ed589", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 694.994158] env[61728]: DEBUG oslo_concurrency.lockutils [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] Acquired lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.994348] env[61728]: DEBUG nova.network.neutron [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Refreshing network info cache for port 124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 694.995607] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:47:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48512b02-ad5c-4105-ba7d-fd4775acf8e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '124d4780-15ec-4f17-a2c4-edf1517ed589', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 695.006023] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Creating folder: Project (01839050e56e4dbfa6e22471a3c60998). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 695.013472] env[61728]: DEBUG nova.compute.manager [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 695.021414] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-265cf725-e511-4839-bcef-061bb318826e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.033968] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Created folder: Project (01839050e56e4dbfa6e22471a3c60998) in parent group-v121913. [ 695.033968] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Creating folder: Instances. Parent ref: group-v122051. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 695.034303] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e435f561-bda2-44f1-8455-6f3a6beed816 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.043942] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Created folder: Instances in parent group-v122051. [ 695.044325] env[61728]: DEBUG oslo.service.loopingcall [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 695.044611] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 695.045311] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4a27b39d-2bd7-4565-8de8-51a3857c954f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.073318] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 695.073318] env[61728]: value = "task-463924" [ 695.073318] env[61728]: _type = "Task" [ 695.073318] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.081503] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463924, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.121030] env[61728]: DEBUG nova.compute.manager [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Stashing vm_state: active {{(pid=61728) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 695.167055] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a9775aa-40d7-4104-acc7-83e7fe77f0e7 tempest-AttachInterfacesV270Test-1407082987 tempest-AttachInterfacesV270Test-1407082987-project-member] Lock "4fcc957e-5175-4589-b4da-d8451d92c815" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.422s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.175776] env[61728]: DEBUG nova.virt.hardware [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 695.176072] env[61728]: DEBUG nova.virt.hardware [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 695.176242] env[61728]: DEBUG nova.virt.hardware [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 695.176433] env[61728]: DEBUG nova.virt.hardware [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 695.176585] env[61728]: DEBUG nova.virt.hardware [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 695.176818] env[61728]: DEBUG nova.virt.hardware [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 695.177110] env[61728]: DEBUG nova.virt.hardware [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 695.177303] env[61728]: DEBUG nova.virt.hardware [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 695.177487] env[61728]: DEBUG nova.virt.hardware [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 695.177657] env[61728]: DEBUG nova.virt.hardware [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 695.177838] env[61728]: DEBUG nova.virt.hardware [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 695.178773] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ad20fb-51b7-46a5-9bd8-bc6e7fffbd95 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.195857] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0f5956-a128-49be-aa99-63596d12f307 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.209113] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463920, 'name': ReconfigVM_Task, 'duration_secs': 0.799228} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.220594] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Reconfigured VM instance instance-0000002d to attach disk [datastore1] b563e521-0e0e-45ea-bd5a-056285ea35b9/b563e521-0e0e-45ea-bd5a-056285ea35b9.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 695.220868] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Creating Virtual Disk of size 1048576 KB and adapter type paraVirtual on the data store {{(pid=61728) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1373}} [ 695.222354] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:57:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f65996a3-f865-4492-9377-cd14ec8b3aae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc02ac93-777e-4bcf-8af5-69f162374735', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 695.232451] env[61728]: DEBUG oslo.service.loopingcall [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 695.234559] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CreateVirtualDisk_Task with opID=oslo.vmware-f99f2ef8-1860-42f0-aaf8-d6ac02df41eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.236728] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 695.237164] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd34412b-0ba6-46e0-b680-301fe80c3727 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.264018] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 695.264018] env[61728]: value = "task-463925" [ 695.264018] env[61728]: _type = "Task" [ 695.264018] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.268165] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 695.268165] env[61728]: value = "task-463926" [ 695.268165] env[61728]: _type = "Task" [ 695.268165] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.268469] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463921, 'name': Destroy_Task, 'duration_secs': 0.626659} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.272179] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Destroyed the VM [ 695.272403] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 695.278651] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e4c11a24-0763-4ac5-ad35-031cff6e1944 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.286898] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463926, 'name': CreateVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.290510] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463925, 'name': CreateVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.290986] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 695.290986] env[61728]: value = "task-463927" [ 695.290986] env[61728]: _type = "Task" [ 695.290986] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.303426] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463927, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.449242] env[61728]: DEBUG nova.network.neutron [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Updated VIF entry in instance network info cache for port 124d4780-15ec-4f17-a2c4-edf1517ed589. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 695.449821] env[61728]: DEBUG nova.network.neutron [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Updating instance_info_cache with network_info: [{"id": "124d4780-15ec-4f17-a2c4-edf1517ed589", "address": "fa:16:3e:6d:47:93", "network": {"id": "6a1908cc-f6aa-4dd7-970e-c8a07e1b16ee", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2047209328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "01839050e56e4dbfa6e22471a3c60998", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap124d4780-15", "ovs_interfaceid": "124d4780-15ec-4f17-a2c4-edf1517ed589", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.545898] env[61728]: DEBUG oslo_concurrency.lockutils [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.592435] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463924, 'name': CreateVM_Task, 'duration_secs': 0.373661} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.592633] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 695.593596] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.593596] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.593879] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 695.594355] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a5111a8-02ff-483b-aee5-56f36af9ba36 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.599334] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 695.599334] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b4febe-6e3f-878b-520e-ff116db89702" [ 695.599334] env[61728]: _type = "Task" [ 695.599334] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.610963] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b4febe-6e3f-878b-520e-ff116db89702, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.650918] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.791274] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463925, 'name': CreateVirtualDisk_Task, 'duration_secs': 0.057083} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.791607] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463926, 'name': CreateVM_Task, 'duration_secs': 0.401068} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.796387] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Created Virtual Disk of size 1048576 KB and type thin {{(pid=61728) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1394}} [ 695.796387] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 695.799868] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78d3d44-55aa-40c6-9155-4ef7b81b86d8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.803295] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.809411] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463927, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.838194] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Reconfiguring VM instance instance-0000002d to attach disk [datastore1] b563e521-0e0e-45ea-bd5a-056285ea35b9/ephemeral_0.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 695.841749] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d388cdcd-2f12-48a7-9243-8af82bd699c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.858111] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Acquiring lock "39eaa6f4-df3f-4928-a1f1-8c861a392e29" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.858422] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Lock "39eaa6f4-df3f-4928-a1f1-8c861a392e29" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.865593] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 695.865593] env[61728]: value = "task-463928" [ 695.865593] env[61728]: _type = "Task" [ 695.865593] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.877039] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463928, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.896838] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a09face-1ebd-4b11-9c6d-1628b97b76b3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.904695] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b99b12-c444-492c-bb9e-86857d7b7e1b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.936581] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf8ac51-91be-44c8-a4a3-b472857e1621 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.948525] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08122da1-54a5-457a-9f68-2b82adb11964 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.954600] env[61728]: DEBUG oslo_concurrency.lockutils [req-bb45b721-ce74-4be0-928d-0f3e24e4b28f req-56a1a074-1c8b-46b0-a9be-efea8ee695b9 service nova] Releasing lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.963357] env[61728]: DEBUG nova.compute.provider_tree [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 696.113163] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b4febe-6e3f-878b-520e-ff116db89702, 'name': SearchDatastore_Task, 'duration_secs': 0.013367} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.113493] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.113736] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 696.113983] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.114232] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.114374] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 696.114618] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.114940] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 696.115202] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2792198-c457-48f4-8167-c302b980e7f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.119910] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1bf9d05-609b-4a17-b33f-88d734eb09fc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.125572] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 696.125572] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5214b8c1-ad3a-b811-679a-29bf2f563195" [ 696.125572] env[61728]: _type = "Task" [ 696.125572] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.130332] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 696.130524] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 696.131774] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-debd4eb5-6d2e-4d19-8fde-4aef8ca110d8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.137361] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5214b8c1-ad3a-b811-679a-29bf2f563195, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.140544] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 696.140544] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb5b56-9c92-bc7c-4e32-645b4ad7071f" [ 696.140544] env[61728]: _type = "Task" [ 696.140544] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.148145] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb5b56-9c92-bc7c-4e32-645b4ad7071f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.304750] env[61728]: DEBUG oslo_vmware.api [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463927, 'name': RemoveSnapshot_Task, 'duration_secs': 0.707689} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.305063] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 696.305311] env[61728]: INFO nova.compute.manager [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Took 16.10 seconds to snapshot the instance on the hypervisor. [ 696.375505] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463928, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.488770] env[61728]: ERROR nova.scheduler.client.report [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] [req-fb2ea380-dcc7-456b-a815-d4693304a538] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-fb2ea380-dcc7-456b-a815-d4693304a538"}]} [ 696.516220] env[61728]: DEBUG nova.scheduler.client.report [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 696.541903] env[61728]: DEBUG nova.scheduler.client.report [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 696.542154] env[61728]: DEBUG nova.compute.provider_tree [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 696.563039] env[61728]: DEBUG nova.scheduler.client.report [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 696.585793] env[61728]: DEBUG nova.scheduler.client.report [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 696.637519] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5214b8c1-ad3a-b811-679a-29bf2f563195, 'name': SearchDatastore_Task, 'duration_secs': 0.032964} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.642460] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.642460] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 696.642460] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.650574] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb5b56-9c92-bc7c-4e32-645b4ad7071f, 'name': SearchDatastore_Task, 'duration_secs': 0.028247} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.651433] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc6d4da9-c595-4220-a2ea-2783987d0a84 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.657137] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 696.657137] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5216d907-30d4-97ad-3587-31283f8f0e2b" [ 696.657137] env[61728]: _type = "Task" [ 696.657137] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.667807] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5216d907-30d4-97ad-3587-31283f8f0e2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.785135] env[61728]: DEBUG nova.compute.manager [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 696.785135] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659ea897-3625-41ab-843d-0e49ec8f403e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.879889] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463928, 'name': ReconfigVM_Task, 'duration_secs': 0.940346} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.886671] env[61728]: DEBUG nova.compute.manager [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Found 3 images (rotation: 2) {{(pid=61728) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 696.889077] env[61728]: DEBUG nova.compute.manager [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Rotating out 1 backups {{(pid=61728) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 696.889077] env[61728]: DEBUG nova.compute.manager [None req-cf4f1282-9d08-4a03-aee8-da202da0e02b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Deleting image 4d578dcb-5948-4a97-a32f-c1442844b66f {{(pid=61728) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 696.889077] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Reconfigured VM instance instance-0000002d to attach disk [datastore1] b563e521-0e0e-45ea-bd5a-056285ea35b9/ephemeral_0.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 696.890080] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a37b060d-0608-4b55-9736-9f17fa1d4b49 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.896044] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 696.896044] env[61728]: value = "task-463929" [ 696.896044] env[61728]: _type = "Task" [ 696.896044] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.906820] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463929, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.171898] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5216d907-30d4-97ad-3587-31283f8f0e2b, 'name': SearchDatastore_Task, 'duration_secs': 0.009549} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.171898] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.171898] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3bfefcfc-db97-4a9d-86cb-9fb1d8158863/3bfefcfc-db97-4a9d-86cb-9fb1d8158863.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 697.171898] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.171898] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 697.172307] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac8a5014-8552-4718-b8dd-7de034b41fa4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.178039] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0d0e88c-b312-42b4-a03d-f04c3251bca4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.186990] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 697.186990] env[61728]: value = "task-463930" [ 697.186990] env[61728]: _type = "Task" [ 697.186990] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.188051] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 697.188420] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 697.195030] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-873d9678-d41b-4e62-9f4d-a46a0410d901 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.207999] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-463930, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.209957] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 697.209957] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e9c833-f416-ee44-64c8-182116fc4323" [ 697.209957] env[61728]: _type = "Task" [ 697.209957] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.221603] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e9c833-f416-ee44-64c8-182116fc4323, 'name': SearchDatastore_Task, 'duration_secs': 0.008513} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.222897] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49eafec0-bcb5-4f23-ae75-85919c3d2a6c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.228411] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 697.228411] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ca545e-9cdc-979f-fe39-e93b83b51547" [ 697.228411] env[61728]: _type = "Task" [ 697.228411] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.238454] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ca545e-9cdc-979f-fe39-e93b83b51547, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.256268] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4964b9-dd0c-43d2-8e39-720d450d00f0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.264884] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-976cf4d5-bb4e-43fa-9f7d-42fabcb9d220 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.297238] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a068bf5e-04d2-4d35-a2e7-256d85f84b70 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.300804] env[61728]: INFO nova.compute.manager [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] instance snapshotting [ 697.306037] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2168f1fb-4e83-4cb1-9e17-1f9f316fe9a0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.310243] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a30f2f-d18f-4781-b46e-f09f8ff0921f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.327661] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b9e337-554d-4c91-9c08-ecbfd4e8f245 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.338880] env[61728]: DEBUG nova.compute.provider_tree [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 697.409476] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463929, 'name': Rename_Task, 'duration_secs': 0.350468} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.409787] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 697.410124] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9e05bba0-5ccd-412e-8ba7-6e7b9b961996 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.418073] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 697.418073] env[61728]: value = "task-463931" [ 697.418073] env[61728]: _type = "Task" [ 697.418073] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.426428] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463931, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.699113] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-463930, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474427} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.699403] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3bfefcfc-db97-4a9d-86cb-9fb1d8158863/3bfefcfc-db97-4a9d-86cb-9fb1d8158863.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 697.700196] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 697.700196] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b46c2b46-db0d-4c97-8995-8a1113a1d8f1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.706011] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 697.706011] env[61728]: value = "task-463932" [ 697.706011] env[61728]: _type = "Task" [ 697.706011] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.715493] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-463932, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.738624] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ca545e-9cdc-979f-fe39-e93b83b51547, 'name': SearchDatastore_Task, 'duration_secs': 0.011098} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.738964] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.739151] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4/771341ed-8b8e-470c-9686-82650f5271b4.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 697.739420] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4748c7cb-447f-4126-9183-e33bc1c0a1eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.746459] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 697.746459] env[61728]: value = "task-463933" [ 697.746459] env[61728]: _type = "Task" [ 697.746459] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.755599] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463933, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.849934] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 697.849934] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2c4bbc85-9585-4acd-952d-16fd1b885769 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.857713] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 697.857713] env[61728]: value = "task-463934" [ 697.857713] env[61728]: _type = "Task" [ 697.857713] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.869119] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463934, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.895718] env[61728]: DEBUG nova.scheduler.client.report [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 77 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 697.895962] env[61728]: DEBUG nova.compute.provider_tree [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 77 to 78 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 697.896181] env[61728]: DEBUG nova.compute.provider_tree [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 697.936089] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463931, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.216802] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-463932, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066579} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.217112] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 698.218016] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a17900c-b935-40d6-a8af-fdb564a2c6cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.242886] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] 3bfefcfc-db97-4a9d-86cb-9fb1d8158863/3bfefcfc-db97-4a9d-86cb-9fb1d8158863.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 698.243597] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a88764bf-dfeb-484c-bd63-2545f2215db8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.273131] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463933, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475165} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.274436] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4/771341ed-8b8e-470c-9686-82650f5271b4.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 698.274668] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 698.274990] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 698.274990] env[61728]: value = "task-463935" [ 698.274990] env[61728]: _type = "Task" [ 698.274990] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.275204] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-29f8cfe0-dc27-4b6b-aedb-87d41774d7cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.285356] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-463935, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.287126] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 698.287126] env[61728]: value = "task-463936" [ 698.287126] env[61728]: _type = "Task" [ 698.287126] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.296755] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463936, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.368107] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463934, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.402380] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.784s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.404875] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.465s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.406533] env[61728]: INFO nova.compute.claims [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 698.429876] env[61728]: DEBUG oslo_vmware.api [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-463931, 'name': PowerOnVM_Task, 'duration_secs': 0.657512} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.430493] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 698.430932] env[61728]: INFO nova.compute.manager [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Took 10.88 seconds to spawn the instance on the hypervisor. [ 698.433017] env[61728]: DEBUG nova.compute.manager [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 698.433017] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29082d2b-1626-4157-82e5-6d863ed7eb84 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.437469] env[61728]: INFO nova.scheduler.client.report [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Deleted allocations for instance 69d7af16-7708-4df6-acca-093b6c1df1ba [ 698.792391] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-463935, 'name': ReconfigVM_Task, 'duration_secs': 0.283331} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.799027] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Reconfigured VM instance instance-0000002e to attach disk [datastore1] 3bfefcfc-db97-4a9d-86cb-9fb1d8158863/3bfefcfc-db97-4a9d-86cb-9fb1d8158863.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 698.799491] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b04c3526-4e4c-4739-b095-70e4f37462e9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.810952] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463936, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062969} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.813434] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 698.814251] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 698.814251] env[61728]: value = "task-463937" [ 698.814251] env[61728]: _type = "Task" [ 698.814251] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.816272] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4624e4c-1ad3-4210-ab4e-3cd483f39ecf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.835839] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-463937, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.861988] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Reconfiguring VM instance instance-00000004 to attach disk [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4/771341ed-8b8e-470c-9686-82650f5271b4.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 698.862921] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c4e6828-dab3-4a57-801a-0e61492edde8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.886752] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463934, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.889029] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 698.889029] env[61728]: value = "task-463938" [ 698.889029] env[61728]: _type = "Task" [ 698.889029] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.900245] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463938, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.959262] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54ef351d-2839-4020-9a0f-0c572b2210c3 tempest-ServersNegativeTestMultiTenantJSON-1431280258 tempest-ServersNegativeTestMultiTenantJSON-1431280258-project-member] Lock "69d7af16-7708-4df6-acca-093b6c1df1ba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.423s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.966464] env[61728]: INFO nova.compute.manager [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Took 46.78 seconds to build instance. [ 699.332233] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-463937, 'name': Rename_Task, 'duration_secs': 0.163941} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.332759] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 699.333162] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-300bcce9-9384-405a-af52-2ba364bc9c8a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.339201] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 699.339201] env[61728]: value = "task-463939" [ 699.339201] env[61728]: _type = "Task" [ 699.339201] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.350129] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-463939, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.374677] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463934, 'name': CreateSnapshot_Task, 'duration_secs': 1.248465} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.378020] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 699.378020] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b11d1a2-a7e4-4a95-a563-d94e78404dda {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.400514] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463938, 'name': ReconfigVM_Task, 'duration_secs': 0.312196} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.400817] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Reconfigured VM instance instance-00000004 to attach disk [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4/771341ed-8b8e-470c-9686-82650f5271b4.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 699.401465] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-83361ee9-a128-4187-b864-b6b87f3a275b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.407210] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 699.407210] env[61728]: value = "task-463940" [ 699.407210] env[61728]: _type = "Task" [ 699.407210] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.420740] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463940, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.472445] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6d0008e3-6a05-441a-b136-9b46a8c3a554 tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "b563e521-0e0e-45ea-bd5a-056285ea35b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.731s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.854741] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-463939, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.903739] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 699.904098] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1dd65f70-412d-4abc-891b-fef3e7f0c1ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.918602] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463940, 'name': Rename_Task, 'duration_secs': 0.173693} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.920274] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 699.920664] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 699.920664] env[61728]: value = "task-463941" [ 699.920664] env[61728]: _type = "Task" [ 699.920664] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.921023] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6c3e5cd-24c9-4f83-b326-c19f79cd5931 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.933124] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463941, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.935164] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 699.935164] env[61728]: value = "task-463942" [ 699.935164] env[61728]: _type = "Task" [ 699.935164] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.941675] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463942, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.979208] env[61728]: DEBUG nova.compute.manager [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 700.042806] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0abcc077-3804-4cdb-b2ca-4391c0c06615 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.054195] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee26a339-13d0-4a9a-9ff9-5b294b7c5ab0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.094832] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a403d81-0662-42f1-b66a-4c0068e537cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.106407] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197ef2c5-be58-49b7-b038-3ffbb4e3fced {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.130870] env[61728]: DEBUG nova.compute.provider_tree [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 700.357631] env[61728]: DEBUG oslo_vmware.api [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-463939, 'name': PowerOnVM_Task, 'duration_secs': 0.52278} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.357631] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 700.357631] env[61728]: INFO nova.compute.manager [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Took 8.23 seconds to spawn the instance on the hypervisor. [ 700.357631] env[61728]: DEBUG nova.compute.manager [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 700.357967] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9296bbde-cc36-449f-8b53-c71b2ba461f0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.434410] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463941, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.448857] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463942, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.496981] env[61728]: DEBUG oslo_concurrency.lockutils [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.680159] env[61728]: DEBUG nova.scheduler.client.report [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 78 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 700.680159] env[61728]: DEBUG nova.compute.provider_tree [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 78 to 79 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 700.680159] env[61728]: DEBUG nova.compute.provider_tree [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 700.893045] env[61728]: INFO nova.compute.manager [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Took 44.07 seconds to build instance. [ 700.938526] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463941, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.950643] env[61728]: DEBUG oslo_vmware.api [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463942, 'name': PowerOnVM_Task, 'duration_secs': 0.880417} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.952019] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 700.952019] env[61728]: DEBUG nova.compute.manager [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 700.952168] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0fec20f-e7b3-4386-a490-ca9d92b3cae6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.187698] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.781s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.188373] env[61728]: DEBUG nova.compute.manager [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 701.192797] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.981s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.195097] env[61728]: INFO nova.compute.claims [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.399298] env[61728]: DEBUG oslo_concurrency.lockutils [None req-54f45302-d8b4-4cb6-bf11-c7924a09dc92 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.017s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.438361] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463941, 'name': CloneVM_Task} progress is 95%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.476839] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.708012] env[61728]: DEBUG nova.compute.utils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 701.710083] env[61728]: DEBUG nova.compute.manager [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 701.710774] env[61728]: DEBUG nova.network.neutron [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 701.800430] env[61728]: DEBUG nova.policy [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '091abd20ebfe450b9283f5c7df1c7993', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c78292171414530a1caba3dfd0f620f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 701.904636] env[61728]: DEBUG nova.compute.manager [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 701.942941] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463941, 'name': CloneVM_Task, 'duration_secs': 1.82669} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.943294] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Created linked-clone VM from snapshot [ 701.947413] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af467b89-d1db-41f8-a065-8348f3dc371f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.963565] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Uploading image 83d104d0-cc75-41db-87ff-ea8445c5e492 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 701.984300] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 701.984924] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b0f5f9f0-30c7-4e9a-9873-546afbd38533 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.997245] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 701.997245] env[61728]: value = "task-463943" [ 701.997245] env[61728]: _type = "Task" [ 701.997245] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.010600] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463943, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.222732] env[61728]: DEBUG nova.compute.manager [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 702.237234] env[61728]: DEBUG nova.compute.manager [req-939fb857-261d-4813-8e9b-d576268cdaaa req-ebb531ad-dee2-4e06-a011-52dfe8383855 service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Received event network-changed-613976dd-b5d2-4919-9887-b3d42dc84c87 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 702.237234] env[61728]: DEBUG nova.compute.manager [req-939fb857-261d-4813-8e9b-d576268cdaaa req-ebb531ad-dee2-4e06-a011-52dfe8383855 service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Refreshing instance network info cache due to event network-changed-613976dd-b5d2-4919-9887-b3d42dc84c87. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 702.237234] env[61728]: DEBUG oslo_concurrency.lockutils [req-939fb857-261d-4813-8e9b-d576268cdaaa req-ebb531ad-dee2-4e06-a011-52dfe8383855 service nova] Acquiring lock "refresh_cache-b563e521-0e0e-45ea-bd5a-056285ea35b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.237432] env[61728]: DEBUG oslo_concurrency.lockutils [req-939fb857-261d-4813-8e9b-d576268cdaaa req-ebb531ad-dee2-4e06-a011-52dfe8383855 service nova] Acquired lock "refresh_cache-b563e521-0e0e-45ea-bd5a-056285ea35b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.237603] env[61728]: DEBUG nova.network.neutron [req-939fb857-261d-4813-8e9b-d576268cdaaa req-ebb531ad-dee2-4e06-a011-52dfe8383855 service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Refreshing network info cache for port 613976dd-b5d2-4919-9887-b3d42dc84c87 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 702.414244] env[61728]: DEBUG nova.network.neutron [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Successfully created port: fa8afb65-af92-44ff-932f-b6e166dc8609 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.434640] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.518291] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463943, 'name': Destroy_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.876971] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "b63fa3df-317a-46d4-b8e9-74e9e287efde" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.876971] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "b63fa3df-317a-46d4-b8e9-74e9e287efde" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.964780] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05f8fbf-d149-47c7-b867-4c3c3baa4012 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.974357] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579d8654-ba62-46fc-916f-b0c29b5f7217 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.014088] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a17d02-f282-44c2-a7b8-8702a48f64b2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.026863] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463943, 'name': Destroy_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.026863] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c7d0965-972d-4e3c-a0c7-a7a9b41c2033 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.045034] env[61728]: DEBUG nova.compute.provider_tree [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.239994] env[61728]: DEBUG nova.compute.manager [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 703.263112] env[61728]: DEBUG nova.network.neutron [req-939fb857-261d-4813-8e9b-d576268cdaaa req-ebb531ad-dee2-4e06-a011-52dfe8383855 service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Updated VIF entry in instance network info cache for port 613976dd-b5d2-4919-9887-b3d42dc84c87. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 703.264062] env[61728]: DEBUG nova.network.neutron [req-939fb857-261d-4813-8e9b-d576268cdaaa req-ebb531ad-dee2-4e06-a011-52dfe8383855 service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Updating instance_info_cache with network_info: [{"id": "613976dd-b5d2-4919-9887-b3d42dc84c87", "address": "fa:16:3e:9b:9b:90", "network": {"id": "113d3399-4dbf-44c8-9fef-d29cbb233c49", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471418633-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab547a56a1e24b3bb051b7f0c70b4640", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap613976dd-b5", "ovs_interfaceid": "613976dd-b5d2-4919-9887-b3d42dc84c87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.268491] env[61728]: DEBUG nova.virt.hardware [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 703.268962] env[61728]: DEBUG nova.virt.hardware [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 703.269116] env[61728]: DEBUG nova.virt.hardware [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.269388] env[61728]: DEBUG nova.virt.hardware [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 703.269813] env[61728]: DEBUG nova.virt.hardware [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.270096] env[61728]: DEBUG nova.virt.hardware [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 703.270386] env[61728]: DEBUG nova.virt.hardware [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 703.270776] env[61728]: DEBUG nova.virt.hardware [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 703.271054] env[61728]: DEBUG nova.virt.hardware [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 703.271446] env[61728]: DEBUG nova.virt.hardware [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 703.271885] env[61728]: DEBUG nova.virt.hardware [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 703.273165] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df96b90-a882-4108-a744-a46e09651762 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.285499] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be04ed0a-7364-46f8-a2f4-b0b91056a16b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.521308] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463943, 'name': Destroy_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.550247] env[61728]: DEBUG nova.scheduler.client.report [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.605799] env[61728]: INFO nova.compute.manager [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Rebuilding instance [ 703.657917] env[61728]: DEBUG nova.compute.manager [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 703.659038] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a47ab0-be49-4346-a78b-6989d8306508 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.767050] env[61728]: DEBUG oslo_concurrency.lockutils [req-939fb857-261d-4813-8e9b-d576268cdaaa req-ebb531ad-dee2-4e06-a011-52dfe8383855 service nova] Releasing lock "refresh_cache-b563e521-0e0e-45ea-bd5a-056285ea35b9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.031128] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463943, 'name': Destroy_Task, 'duration_secs': 1.738715} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.031128] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Destroyed the VM [ 704.031128] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 704.031128] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-36e69fa3-cf88-45dd-b785-6eb1f888c718 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.040050] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 704.040050] env[61728]: value = "task-463944" [ 704.040050] env[61728]: _type = "Task" [ 704.040050] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.055282] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463944, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.060086] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.867s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.060632] env[61728]: DEBUG nova.compute.manager [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 704.069142] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.380s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.073041] env[61728]: INFO nova.compute.claims [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 704.174064] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 704.174323] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa826b5f-aeee-43c8-897d-d370331cf3d0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.185023] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 704.185023] env[61728]: value = "task-463945" [ 704.185023] env[61728]: _type = "Task" [ 704.185023] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.195189] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463945, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.555501] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463944, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.575065] env[61728]: DEBUG nova.compute.utils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.576550] env[61728]: DEBUG nova.compute.manager [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.576937] env[61728]: DEBUG nova.network.neutron [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 704.597686] env[61728]: DEBUG nova.network.neutron [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Successfully updated port: fa8afb65-af92-44ff-932f-b6e166dc8609 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 704.667308] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a10a8a9b-4cdd-4f64-b0e4-5e7038c57886 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.667308] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a10a8a9b-4cdd-4f64-b0e4-5e7038c57886 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.667308] env[61728]: DEBUG nova.compute.manager [None req-a10a8a9b-4cdd-4f64-b0e4-5e7038c57886 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 704.667308] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5aebbb6-57c7-4538-8b95-f44c5aed5991 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.677891] env[61728]: DEBUG nova.compute.manager [None req-a10a8a9b-4cdd-4f64-b0e4-5e7038c57886 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61728) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 704.678535] env[61728]: DEBUG nova.objects.instance [None req-a10a8a9b-4cdd-4f64-b0e4-5e7038c57886 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'flavor' on Instance uuid 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 704.699645] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463945, 'name': PowerOffVM_Task, 'duration_secs': 0.336724} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.699793] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 704.700277] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 704.700804] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e67bee-d8e9-48eb-9038-75c065df6ac1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.705030] env[61728]: DEBUG nova.policy [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8a56b87b2294bc8b71a599c076414d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8fc16f1d1b144286b180cd7c8f7588fd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 704.712727] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 704.712727] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fdd21710-7e05-4fcc-91ea-79567a77ee46 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.788071] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 704.789736] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 704.789736] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleting the datastore file [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 704.789736] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc0604df-fc2f-4efa-a92d-e00af784c773 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.798566] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 704.798566] env[61728]: value = "task-463947" [ 704.798566] env[61728]: _type = "Task" [ 704.798566] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.808909] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463947, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.827915] env[61728]: DEBUG nova.compute.manager [req-0b818692-0242-4b07-90ef-79a54ab608f1 req-935365a2-2e19-405e-8c5f-d46106665831 service nova] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Received event network-vif-plugged-fa8afb65-af92-44ff-932f-b6e166dc8609 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 704.828520] env[61728]: DEBUG oslo_concurrency.lockutils [req-0b818692-0242-4b07-90ef-79a54ab608f1 req-935365a2-2e19-405e-8c5f-d46106665831 service nova] Acquiring lock "c14034b5-330b-4fb1-af31-d530e3231992-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.829121] env[61728]: DEBUG oslo_concurrency.lockutils [req-0b818692-0242-4b07-90ef-79a54ab608f1 req-935365a2-2e19-405e-8c5f-d46106665831 service nova] Lock "c14034b5-330b-4fb1-af31-d530e3231992-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.829121] env[61728]: DEBUG oslo_concurrency.lockutils [req-0b818692-0242-4b07-90ef-79a54ab608f1 req-935365a2-2e19-405e-8c5f-d46106665831 service nova] Lock "c14034b5-330b-4fb1-af31-d530e3231992-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.829305] env[61728]: DEBUG nova.compute.manager [req-0b818692-0242-4b07-90ef-79a54ab608f1 req-935365a2-2e19-405e-8c5f-d46106665831 service nova] [instance: c14034b5-330b-4fb1-af31-d530e3231992] No waiting events found dispatching network-vif-plugged-fa8afb65-af92-44ff-932f-b6e166dc8609 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 704.829645] env[61728]: WARNING nova.compute.manager [req-0b818692-0242-4b07-90ef-79a54ab608f1 req-935365a2-2e19-405e-8c5f-d46106665831 service nova] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Received unexpected event network-vif-plugged-fa8afb65-af92-44ff-932f-b6e166dc8609 for instance with vm_state building and task_state spawning. [ 705.037152] env[61728]: DEBUG nova.network.neutron [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Successfully created port: 401726ae-678d-4918-b8c1-441b3802ddc8 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.052039] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463944, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.080931] env[61728]: DEBUG nova.compute.manager [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 705.100943] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "refresh_cache-c14034b5-330b-4fb1-af31-d530e3231992" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.101541] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "refresh_cache-c14034b5-330b-4fb1-af31-d530e3231992" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.101541] env[61728]: DEBUG nova.network.neutron [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 705.192691] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a10a8a9b-4cdd-4f64-b0e4-5e7038c57886 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 705.192691] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2697edaa-ef98-4ee0-bbaf-e2f87daccee2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.203200] env[61728]: DEBUG oslo_vmware.api [None req-a10a8a9b-4cdd-4f64-b0e4-5e7038c57886 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 705.203200] env[61728]: value = "task-463948" [ 705.203200] env[61728]: _type = "Task" [ 705.203200] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.219346] env[61728]: DEBUG oslo_vmware.api [None req-a10a8a9b-4cdd-4f64-b0e4-5e7038c57886 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463948, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.225531] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.226501] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.310956] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463947, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.343846} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.311375] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 705.311495] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 705.311689] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 705.554251] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463944, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.660685] env[61728]: DEBUG nova.network.neutron [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.714580] env[61728]: DEBUG oslo_vmware.api [None req-a10a8a9b-4cdd-4f64-b0e4-5e7038c57886 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-463948, 'name': PowerOffVM_Task, 'duration_secs': 0.228545} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.714847] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a10a8a9b-4cdd-4f64-b0e4-5e7038c57886 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 705.715038] env[61728]: DEBUG nova.compute.manager [None req-a10a8a9b-4cdd-4f64-b0e4-5e7038c57886 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 705.715833] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166abb06-19a3-4888-988e-18bfd72081dc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.737167] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c081cb0-23ef-430c-b702-ac9b7843cfb6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.748810] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8229fd5a-a43c-4f80-8808-cf54760224c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.789896] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6caffccb-9130-4563-8587-7418128b87d0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.802288] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5d9bb4b-523c-4cea-975b-1ebed5ee893c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.828020] env[61728]: DEBUG nova.compute.provider_tree [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.935182] env[61728]: DEBUG nova.network.neutron [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Updating instance_info_cache with network_info: [{"id": "fa8afb65-af92-44ff-932f-b6e166dc8609", "address": "fa:16:3e:11:b1:a4", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa8afb65-af", "ovs_interfaceid": "fa8afb65-af92-44ff-932f-b6e166dc8609", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.985419] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Acquiring lock "a5c7e3ff-6829-4528-9279-e865a0eb8512" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.986034] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Lock "a5c7e3ff-6829-4528-9279-e865a0eb8512" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.055750] env[61728]: DEBUG oslo_vmware.api [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463944, 'name': RemoveSnapshot_Task, 'duration_secs': 1.978665} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.056591] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 706.093558] env[61728]: DEBUG nova.compute.manager [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 706.124722] env[61728]: DEBUG nova.virt.hardware [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 706.125017] env[61728]: DEBUG nova.virt.hardware [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 706.125198] env[61728]: DEBUG nova.virt.hardware [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.125402] env[61728]: DEBUG nova.virt.hardware [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 706.125571] env[61728]: DEBUG nova.virt.hardware [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.125736] env[61728]: DEBUG nova.virt.hardware [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 706.125982] env[61728]: DEBUG nova.virt.hardware [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 706.126260] env[61728]: DEBUG nova.virt.hardware [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 706.126480] env[61728]: DEBUG nova.virt.hardware [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 706.126675] env[61728]: DEBUG nova.virt.hardware [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 706.126870] env[61728]: DEBUG nova.virt.hardware [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.127877] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-252ec950-d7b7-4a15-9720-c9de3743bbd3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.140018] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8632fd-47ba-4cee-ba5f-5dacbc4f7922 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.238201] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a10a8a9b-4cdd-4f64-b0e4-5e7038c57886 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.572s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.331660] env[61728]: DEBUG nova.scheduler.client.report [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.361853] env[61728]: DEBUG nova.virt.hardware [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 706.361853] env[61728]: DEBUG nova.virt.hardware [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 706.361853] env[61728]: DEBUG nova.virt.hardware [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.361853] env[61728]: DEBUG nova.virt.hardware [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 706.364050] env[61728]: DEBUG nova.virt.hardware [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.364283] env[61728]: DEBUG nova.virt.hardware [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 706.364513] env[61728]: DEBUG nova.virt.hardware [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 706.364683] env[61728]: DEBUG nova.virt.hardware [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 706.366083] env[61728]: DEBUG nova.virt.hardware [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 706.366281] env[61728]: DEBUG nova.virt.hardware [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 706.366485] env[61728]: DEBUG nova.virt.hardware [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.368424] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c8e0f7-a8d2-4733-b9e0-a9c65e86c431 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.378353] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd469b39-5d99-4a93-a362-d9c09ac5b3ab {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.394489] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:57:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f65996a3-f865-4492-9377-cd14ec8b3aae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc02ac93-777e-4bcf-8af5-69f162374735', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 706.402146] env[61728]: DEBUG oslo.service.loopingcall [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 706.402432] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 706.402656] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-18cd8e02-e529-433f-a915-1c1633eec401 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.428953] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 706.428953] env[61728]: value = "task-463949" [ 706.428953] env[61728]: _type = "Task" [ 706.428953] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.439033] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "refresh_cache-c14034b5-330b-4fb1-af31-d530e3231992" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.439786] env[61728]: DEBUG nova.compute.manager [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Instance network_info: |[{"id": "fa8afb65-af92-44ff-932f-b6e166dc8609", "address": "fa:16:3e:11:b1:a4", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa8afb65-af", "ovs_interfaceid": "fa8afb65-af92-44ff-932f-b6e166dc8609", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 706.439786] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463949, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.440127] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:b1:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fa8afb65-af92-44ff-932f-b6e166dc8609', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 706.448093] env[61728]: DEBUG oslo.service.loopingcall [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 706.448637] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 706.448878] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3ca9611b-3196-4b03-96b2-881ad47c12c5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.471954] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 706.471954] env[61728]: value = "task-463950" [ 706.471954] env[61728]: _type = "Task" [ 706.471954] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.481637] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463950, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.562781] env[61728]: WARNING nova.compute.manager [None req-f8549a83-23e2-43cc-bd8d-57d4d21f2ee6 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Image not found during snapshot: nova.exception.ImageNotFound: Image 83d104d0-cc75-41db-87ff-ea8445c5e492 could not be found. [ 706.839121] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.770s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.839838] env[61728]: DEBUG nova.compute.manager [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 706.843405] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.656s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.845296] env[61728]: INFO nova.compute.claims [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.924156] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "ca8d41dc-5719-4ade-b82a-b1d7f1c8def1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.924446] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "ca8d41dc-5719-4ade-b82a-b1d7f1c8def1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.924668] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "ca8d41dc-5719-4ade-b82a-b1d7f1c8def1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.924856] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "ca8d41dc-5719-4ade-b82a-b1d7f1c8def1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.925048] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "ca8d41dc-5719-4ade-b82a-b1d7f1c8def1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.927808] env[61728]: INFO nova.compute.manager [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Terminating instance [ 706.937018] env[61728]: DEBUG nova.compute.manager [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 706.937018] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 706.937018] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a342dc38-7126-4589-9961-472e700e60ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.945930] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463949, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.948437] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 706.949150] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f1ac9057-c794-4103-93a6-5840dba63c28 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.957353] env[61728]: DEBUG oslo_vmware.api [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 706.957353] env[61728]: value = "task-463951" [ 706.957353] env[61728]: _type = "Task" [ 706.957353] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.971157] env[61728]: DEBUG oslo_vmware.api [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463951, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.982096] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463950, 'name': CreateVM_Task, 'duration_secs': 0.457377} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.982500] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 706.983334] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.983749] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.984161] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 706.984539] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03813481-918d-419f-bff7-1a3fc44ce567 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.987499] env[61728]: DEBUG nova.network.neutron [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Successfully updated port: 401726ae-678d-4918-b8c1-441b3802ddc8 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 706.993525] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 706.993525] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52614d41-bd40-c663-6913-e032ebae100d" [ 706.993525] env[61728]: _type = "Task" [ 706.993525] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.002937] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52614d41-bd40-c663-6913-e032ebae100d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.154709] env[61728]: DEBUG nova.compute.manager [req-64bb4d6a-faa9-4df2-9376-7f13288504da req-14bbc11a-2bce-4213-a841-ad8c525e5e61 service nova] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Received event network-changed-fa8afb65-af92-44ff-932f-b6e166dc8609 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 707.154965] env[61728]: DEBUG nova.compute.manager [req-64bb4d6a-faa9-4df2-9376-7f13288504da req-14bbc11a-2bce-4213-a841-ad8c525e5e61 service nova] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Refreshing instance network info cache due to event network-changed-fa8afb65-af92-44ff-932f-b6e166dc8609. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 707.155213] env[61728]: DEBUG oslo_concurrency.lockutils [req-64bb4d6a-faa9-4df2-9376-7f13288504da req-14bbc11a-2bce-4213-a841-ad8c525e5e61 service nova] Acquiring lock "refresh_cache-c14034b5-330b-4fb1-af31-d530e3231992" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.155362] env[61728]: DEBUG oslo_concurrency.lockutils [req-64bb4d6a-faa9-4df2-9376-7f13288504da req-14bbc11a-2bce-4213-a841-ad8c525e5e61 service nova] Acquired lock "refresh_cache-c14034b5-330b-4fb1-af31-d530e3231992" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.155528] env[61728]: DEBUG nova.network.neutron [req-64bb4d6a-faa9-4df2-9376-7f13288504da req-14bbc11a-2bce-4213-a841-ad8c525e5e61 service nova] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Refreshing network info cache for port fa8afb65-af92-44ff-932f-b6e166dc8609 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 707.353766] env[61728]: DEBUG nova.compute.utils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 707.358385] env[61728]: DEBUG nova.compute.manager [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 707.358555] env[61728]: DEBUG nova.network.neutron [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 707.427418] env[61728]: DEBUG nova.policy [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '296d209aeba746cd8fb37ef4e7fdc86c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '68843e3f2b9a4e2fb9534db4fe5c2ec7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 707.440616] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463949, 'name': CreateVM_Task, 'duration_secs': 0.521961} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.440799] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 707.441690] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.469563] env[61728]: DEBUG oslo_vmware.api [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463951, 'name': PowerOffVM_Task, 'duration_secs': 0.275925} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.469825] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 707.470014] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 707.471361] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c19cf0e-6354-46db-8c38-f8f3392492b3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.492053] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.492053] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquired lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.492053] env[61728]: DEBUG nova.network.neutron [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 707.505664] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52614d41-bd40-c663-6913-e032ebae100d, 'name': SearchDatastore_Task, 'duration_secs': 0.016044} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.506985] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.506985] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 707.506985] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.506985] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.506985] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 707.507472] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.507601] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 707.508349] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7ce70177-7e5f-4a1a-b7bb-694b450ed557 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.510697] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8562643c-23fd-485d-bdf3-047bf19f1988 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.517070] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 707.517070] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52861426-eaac-fafc-6def-0db487bb0b33" [ 707.517070] env[61728]: _type = "Task" [ 707.517070] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.522417] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 707.522467] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 707.527143] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54b8bc14-fa91-4f7a-bc90-66f514a436c8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.529107] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52861426-eaac-fafc-6def-0db487bb0b33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.533694] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 707.533694] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d8449c-a457-163e-f22b-226b33d34b7b" [ 707.533694] env[61728]: _type = "Task" [ 707.533694] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.543214] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d8449c-a457-163e-f22b-226b33d34b7b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.557327] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 707.557556] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 707.557742] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Deleting the datastore file [datastore1] ca8d41dc-5719-4ade-b82a-b1d7f1c8def1 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 707.558037] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f9d1b1bc-c63a-4c12-91c5-d0d8237221fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.568334] env[61728]: DEBUG oslo_vmware.api [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 707.568334] env[61728]: value = "task-463953" [ 707.568334] env[61728]: _type = "Task" [ 707.568334] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.577639] env[61728]: DEBUG oslo_vmware.api [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463953, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.864493] env[61728]: DEBUG nova.compute.manager [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 707.919012] env[61728]: DEBUG nova.network.neutron [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Successfully created port: 8fcb3f51-74e2-4548-8b86-bc5bc9156f40 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 708.026440] env[61728]: DEBUG nova.network.neutron [req-64bb4d6a-faa9-4df2-9376-7f13288504da req-14bbc11a-2bce-4213-a841-ad8c525e5e61 service nova] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Updated VIF entry in instance network info cache for port fa8afb65-af92-44ff-932f-b6e166dc8609. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 708.026854] env[61728]: DEBUG nova.network.neutron [req-64bb4d6a-faa9-4df2-9376-7f13288504da req-14bbc11a-2bce-4213-a841-ad8c525e5e61 service nova] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Updating instance_info_cache with network_info: [{"id": "fa8afb65-af92-44ff-932f-b6e166dc8609", "address": "fa:16:3e:11:b1:a4", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa8afb65-af", "ovs_interfaceid": "fa8afb65-af92-44ff-932f-b6e166dc8609", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.036528] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52861426-eaac-fafc-6def-0db487bb0b33, 'name': SearchDatastore_Task, 'duration_secs': 0.026802} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.046688] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.046895] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 708.047138] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.047973] env[61728]: DEBUG nova.network.neutron [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.063470] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d8449c-a457-163e-f22b-226b33d34b7b, 'name': SearchDatastore_Task, 'duration_secs': 0.025871} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.064261] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9546f7b7-5749-49c0-9d9f-0f609dc9e37e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.075930] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 708.075930] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c4c04f-7ac2-d468-91b2-96297af9ba1e" [ 708.075930] env[61728]: _type = "Task" [ 708.075930] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.085588] env[61728]: DEBUG oslo_vmware.api [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-463953, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.288247} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.086236] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 708.086536] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 708.086781] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 708.086982] env[61728]: INFO nova.compute.manager [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Took 1.15 seconds to destroy the instance on the hypervisor. [ 708.087247] env[61728]: DEBUG oslo.service.loopingcall [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.087844] env[61728]: DEBUG nova.compute.manager [-] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.087844] env[61728]: DEBUG nova.network.neutron [-] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 708.109761] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c4c04f-7ac2-d468-91b2-96297af9ba1e, 'name': SearchDatastore_Task, 'duration_secs': 0.010074} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.112335] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.112596] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] c14034b5-330b-4fb1-af31-d530e3231992/c14034b5-330b-4fb1-af31-d530e3231992.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 708.112934] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.113161] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 708.113401] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6643d0b3-4fbd-444e-8150-c45b7af3f33b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.117495] env[61728]: DEBUG nova.compute.manager [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Stashing vm_state: stopped {{(pid=61728) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 708.119771] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-289c2d95-eb74-4266-8f0d-8371dd2152f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.132034] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 708.132034] env[61728]: value = "task-463954" [ 708.132034] env[61728]: _type = "Task" [ 708.132034] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.132034] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 708.132034] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 708.134619] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-390646ff-e0fe-4d1e-800d-6212f5fff59e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.149234] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463954, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.150471] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 708.150471] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5215e8b7-2f4b-b542-5a94-b606db0bda51" [ 708.150471] env[61728]: _type = "Task" [ 708.150471] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.161905] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5215e8b7-2f4b-b542-5a94-b606db0bda51, 'name': SearchDatastore_Task, 'duration_secs': 0.01046} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.163054] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1651583-6a4e-4139-8375-e1f62957fdcc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.169593] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 708.169593] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52644dc7-b794-d306-c867-9ff52ef2e13c" [ 708.169593] env[61728]: _type = "Task" [ 708.169593] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.178694] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52644dc7-b794-d306-c867-9ff52ef2e13c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.381354] env[61728]: DEBUG nova.network.neutron [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Updating instance_info_cache with network_info: [{"id": "401726ae-678d-4918-b8c1-441b3802ddc8", "address": "fa:16:3e:36:16:cf", "network": {"id": "5d1e643d-f8a4-4302-9920-bb661df4632d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1891316321-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fc16f1d1b144286b180cd7c8f7588fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap401726ae-67", "ovs_interfaceid": "401726ae-678d-4918-b8c1-441b3802ddc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.538105] env[61728]: DEBUG oslo_concurrency.lockutils [req-64bb4d6a-faa9-4df2-9376-7f13288504da req-14bbc11a-2bce-4213-a841-ad8c525e5e61 service nova] Releasing lock "refresh_cache-c14034b5-330b-4fb1-af31-d530e3231992" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.538540] env[61728]: DEBUG nova.compute.manager [req-64bb4d6a-faa9-4df2-9376-7f13288504da req-14bbc11a-2bce-4213-a841-ad8c525e5e61 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Received event network-vif-plugged-401726ae-678d-4918-b8c1-441b3802ddc8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 708.539049] env[61728]: DEBUG oslo_concurrency.lockutils [req-64bb4d6a-faa9-4df2-9376-7f13288504da req-14bbc11a-2bce-4213-a841-ad8c525e5e61 service nova] Acquiring lock "223b1f93-d33e-4566-8c8e-75234a491814-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.539535] env[61728]: DEBUG oslo_concurrency.lockutils [req-64bb4d6a-faa9-4df2-9376-7f13288504da req-14bbc11a-2bce-4213-a841-ad8c525e5e61 service nova] Lock "223b1f93-d33e-4566-8c8e-75234a491814-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.539894] env[61728]: DEBUG oslo_concurrency.lockutils [req-64bb4d6a-faa9-4df2-9376-7f13288504da req-14bbc11a-2bce-4213-a841-ad8c525e5e61 service nova] Lock "223b1f93-d33e-4566-8c8e-75234a491814-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.540237] env[61728]: DEBUG nova.compute.manager [req-64bb4d6a-faa9-4df2-9376-7f13288504da req-14bbc11a-2bce-4213-a841-ad8c525e5e61 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] No waiting events found dispatching network-vif-plugged-401726ae-678d-4918-b8c1-441b3802ddc8 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 708.540566] env[61728]: WARNING nova.compute.manager [req-64bb4d6a-faa9-4df2-9376-7f13288504da req-14bbc11a-2bce-4213-a841-ad8c525e5e61 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Received unexpected event network-vif-plugged-401726ae-678d-4918-b8c1-441b3802ddc8 for instance with vm_state building and task_state spawning. [ 708.578138] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be703ab7-7849-477e-8eaa-184e745b2f13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.589735] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b8341d-2ecc-4cc4-b8f4-e76688319b4c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.628191] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584a2343-ebbd-4be2-9979-d28189377149 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.638259] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d3e6d2-ee87-4929-9dbd-8f7ecb5506f3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.643723] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.648154] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463954, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.659350] env[61728]: DEBUG nova.compute.provider_tree [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.681172] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52644dc7-b794-d306-c867-9ff52ef2e13c, 'name': SearchDatastore_Task, 'duration_secs': 0.010309} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.681719] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.681778] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4/771341ed-8b8e-470c-9686-82650f5271b4.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 708.682039] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c803950-b86d-4f7c-bab7-84d21279fa80 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.690237] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 708.690237] env[61728]: value = "task-463955" [ 708.690237] env[61728]: _type = "Task" [ 708.690237] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.699207] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463955, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.882404] env[61728]: DEBUG nova.compute.manager [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 708.885237] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Releasing lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.885566] env[61728]: DEBUG nova.compute.manager [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Instance network_info: |[{"id": "401726ae-678d-4918-b8c1-441b3802ddc8", "address": "fa:16:3e:36:16:cf", "network": {"id": "5d1e643d-f8a4-4302-9920-bb661df4632d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1891316321-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fc16f1d1b144286b180cd7c8f7588fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap401726ae-67", "ovs_interfaceid": "401726ae-678d-4918-b8c1-441b3802ddc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 708.886261] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:16:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '877ed63d-906e-4bd5-a1fc-7e82d172d41e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '401726ae-678d-4918-b8c1-441b3802ddc8', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 708.893919] env[61728]: DEBUG oslo.service.loopingcall [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.894732] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 708.895412] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-94f725af-c693-400c-9bf2-1c18511f89f6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.921407] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 708.921407] env[61728]: value = "task-463956" [ 708.921407] env[61728]: _type = "Task" [ 708.921407] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.923931] env[61728]: DEBUG nova.virt.hardware [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 708.924198] env[61728]: DEBUG nova.virt.hardware [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 708.924363] env[61728]: DEBUG nova.virt.hardware [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 708.924550] env[61728]: DEBUG nova.virt.hardware [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 708.924702] env[61728]: DEBUG nova.virt.hardware [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 708.924889] env[61728]: DEBUG nova.virt.hardware [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 708.925141] env[61728]: DEBUG nova.virt.hardware [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 708.925319] env[61728]: DEBUG nova.virt.hardware [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 708.925497] env[61728]: DEBUG nova.virt.hardware [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 708.925665] env[61728]: DEBUG nova.virt.hardware [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 708.925846] env[61728]: DEBUG nova.virt.hardware [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 708.926799] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9413b953-d270-49df-9db9-e6185457e250 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.944014] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f46c9b52-13f8-43c3-b5eb-1b029c092174 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.950854] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463956, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.084748] env[61728]: DEBUG nova.network.neutron [-] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.146280] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463954, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521199} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.146666] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] c14034b5-330b-4fb1-af31-d530e3231992/c14034b5-330b-4fb1-af31-d530e3231992.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 709.146961] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 709.147312] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18dc0af1-96c3-4810-ba42-5dee9b5794f7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.159505] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 709.159505] env[61728]: value = "task-463957" [ 709.159505] env[61728]: _type = "Task" [ 709.159505] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.165824] env[61728]: DEBUG nova.scheduler.client.report [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 709.180685] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463957, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.207714] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463955, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.246615] env[61728]: DEBUG oslo_concurrency.lockutils [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Acquiring lock "8c53c9c9-cdc1-407b-8827-8409cf137235" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.246918] env[61728]: DEBUG oslo_concurrency.lockutils [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Lock "8c53c9c9-cdc1-407b-8827-8409cf137235" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.248165] env[61728]: DEBUG oslo_concurrency.lockutils [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Acquiring lock "8c53c9c9-cdc1-407b-8827-8409cf137235-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.248323] env[61728]: DEBUG oslo_concurrency.lockutils [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Lock "8c53c9c9-cdc1-407b-8827-8409cf137235-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.248519] env[61728]: DEBUG oslo_concurrency.lockutils [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Lock "8c53c9c9-cdc1-407b-8827-8409cf137235-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.250774] env[61728]: INFO nova.compute.manager [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Terminating instance [ 709.252667] env[61728]: DEBUG nova.compute.manager [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 709.252875] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 709.253713] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d54795d-04fa-46f7-af5a-c782ca67c53d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.261989] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 709.262614] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1cc8c4e2-c3f7-42ae-b73e-a553899e91f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.270470] env[61728]: DEBUG oslo_vmware.api [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Waiting for the task: (returnval){ [ 709.270470] env[61728]: value = "task-463958" [ 709.270470] env[61728]: _type = "Task" [ 709.270470] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.278408] env[61728]: DEBUG oslo_vmware.api [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463958, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.335300] env[61728]: DEBUG nova.compute.manager [req-714416d9-c3fb-4026-b2e9-53b8dff1035b req-7d3305c3-fee2-47cb-a4e8-9b91fc1b9eb6 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Received event network-changed-401726ae-678d-4918-b8c1-441b3802ddc8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 709.335370] env[61728]: DEBUG nova.compute.manager [req-714416d9-c3fb-4026-b2e9-53b8dff1035b req-7d3305c3-fee2-47cb-a4e8-9b91fc1b9eb6 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Refreshing instance network info cache due to event network-changed-401726ae-678d-4918-b8c1-441b3802ddc8. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 709.337049] env[61728]: DEBUG oslo_concurrency.lockutils [req-714416d9-c3fb-4026-b2e9-53b8dff1035b req-7d3305c3-fee2-47cb-a4e8-9b91fc1b9eb6 service nova] Acquiring lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.337049] env[61728]: DEBUG oslo_concurrency.lockutils [req-714416d9-c3fb-4026-b2e9-53b8dff1035b req-7d3305c3-fee2-47cb-a4e8-9b91fc1b9eb6 service nova] Acquired lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.337049] env[61728]: DEBUG nova.network.neutron [req-714416d9-c3fb-4026-b2e9-53b8dff1035b req-7d3305c3-fee2-47cb-a4e8-9b91fc1b9eb6 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Refreshing network info cache for port 401726ae-678d-4918-b8c1-441b3802ddc8 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 709.438528] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463956, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.587095] env[61728]: INFO nova.compute.manager [-] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Took 1.50 seconds to deallocate network for instance. [ 709.670007] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463957, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084301} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.670315] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 709.671202] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdbb1437-7d70-4088-8ac1-95673608c768 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.674420] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.831s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.675109] env[61728]: DEBUG nova.compute.manager [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 709.677426] env[61728]: DEBUG oslo_concurrency.lockutils [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.508s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.678026] env[61728]: DEBUG nova.objects.instance [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'resources' on Instance uuid 620bb472-c36e-4c56-acdf-42e02614856b {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 709.700924] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] c14034b5-330b-4fb1-af31-d530e3231992/c14034b5-330b-4fb1-af31-d530e3231992.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 709.702027] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ebfa1ca7-e3d6-4bfe-991b-020c6a470f01 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.727684] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463955, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.551065} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.729755] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4/771341ed-8b8e-470c-9686-82650f5271b4.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 709.729755] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 709.730298] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 709.730298] env[61728]: value = "task-463959" [ 709.730298] env[61728]: _type = "Task" [ 709.730298] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.730759] env[61728]: DEBUG nova.network.neutron [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Successfully updated port: 8fcb3f51-74e2-4548-8b86-bc5bc9156f40 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 709.736395] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed48a313-fa15-4ea5-8c30-a77ce34d48c0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.747612] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463959, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.749011] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 709.749011] env[61728]: value = "task-463960" [ 709.749011] env[61728]: _type = "Task" [ 709.749011] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.759515] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463960, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.780514] env[61728]: DEBUG oslo_vmware.api [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463958, 'name': PowerOffVM_Task, 'duration_secs': 0.21517} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.780831] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 709.781024] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 709.781321] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0c7f1b4d-241e-4a2d-8178-e11011bfc86b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.851102] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 709.851470] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 709.851796] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Deleting the datastore file [datastore1] 8c53c9c9-cdc1-407b-8827-8409cf137235 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 709.852116] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ac64a7fd-2168-4a21-b078-7e21dfbc2bce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.860114] env[61728]: DEBUG oslo_vmware.api [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Waiting for the task: (returnval){ [ 709.860114] env[61728]: value = "task-463962" [ 709.860114] env[61728]: _type = "Task" [ 709.860114] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.869339] env[61728]: DEBUG oslo_vmware.api [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463962, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.939020] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463956, 'name': CreateVM_Task, 'duration_secs': 0.890963} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.939203] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 709.939873] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.940059] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.940447] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 709.940717] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c6de652-8df6-46c6-93ef-e60c3287ab58 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.946069] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 709.946069] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d448e0-6d5f-cf50-2e58-a9628fc16696" [ 709.946069] env[61728]: _type = "Task" [ 709.946069] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.954469] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d448e0-6d5f-cf50-2e58-a9628fc16696, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.056599] env[61728]: DEBUG nova.network.neutron [req-714416d9-c3fb-4026-b2e9-53b8dff1035b req-7d3305c3-fee2-47cb-a4e8-9b91fc1b9eb6 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Updated VIF entry in instance network info cache for port 401726ae-678d-4918-b8c1-441b3802ddc8. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 710.056979] env[61728]: DEBUG nova.network.neutron [req-714416d9-c3fb-4026-b2e9-53b8dff1035b req-7d3305c3-fee2-47cb-a4e8-9b91fc1b9eb6 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Updating instance_info_cache with network_info: [{"id": "401726ae-678d-4918-b8c1-441b3802ddc8", "address": "fa:16:3e:36:16:cf", "network": {"id": "5d1e643d-f8a4-4302-9920-bb661df4632d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1891316321-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fc16f1d1b144286b180cd7c8f7588fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap401726ae-67", "ovs_interfaceid": "401726ae-678d-4918-b8c1-441b3802ddc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.094824] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.181554] env[61728]: DEBUG nova.compute.utils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.185647] env[61728]: DEBUG nova.compute.manager [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 710.185845] env[61728]: DEBUG nova.network.neutron [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 710.238820] env[61728]: DEBUG nova.policy [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '296d209aeba746cd8fb37ef4e7fdc86c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '68843e3f2b9a4e2fb9534db4fe5c2ec7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 710.241103] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "refresh_cache-bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.241305] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquired lock "refresh_cache-bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.241500] env[61728]: DEBUG nova.network.neutron [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 710.265454] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463959, 'name': ReconfigVM_Task, 'duration_secs': 0.309245} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.270279] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Reconfigured VM instance instance-0000002f to attach disk [datastore1] c14034b5-330b-4fb1-af31-d530e3231992/c14034b5-330b-4fb1-af31-d530e3231992.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 710.271479] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ddfae429-f6f4-4397-8bb8-1b15811cf0d3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.276490] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463960, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082271} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.277189] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 710.278063] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1394fa-aea8-4b86-ab56-2fb2a0ea58e9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.286646] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 710.286646] env[61728]: value = "task-463963" [ 710.286646] env[61728]: _type = "Task" [ 710.286646] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.309843] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Reconfiguring VM instance instance-00000004 to attach disk [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4/771341ed-8b8e-470c-9686-82650f5271b4.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 710.312145] env[61728]: DEBUG nova.network.neutron [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.317148] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd94fa67-d1b5-4a34-a15e-0e278b96e6dc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.337069] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463963, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.344406] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 710.344406] env[61728]: value = "task-463964" [ 710.344406] env[61728]: _type = "Task" [ 710.344406] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.361801] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463964, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.376126] env[61728]: DEBUG oslo_vmware.api [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Task: {'id': task-463962, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168125} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.377471] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 710.377471] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 710.377471] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 710.377471] env[61728]: INFO nova.compute.manager [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Took 1.12 seconds to destroy the instance on the hypervisor. [ 710.377729] env[61728]: DEBUG oslo.service.loopingcall [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 710.377729] env[61728]: DEBUG nova.compute.manager [-] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.378487] env[61728]: DEBUG nova.network.neutron [-] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 710.458332] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d448e0-6d5f-cf50-2e58-a9628fc16696, 'name': SearchDatastore_Task, 'duration_secs': 0.008935} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.458676] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.459024] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 710.459171] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.459321] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.459507] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 710.459915] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a46cfa54-04e4-43ad-8aca-e9cece4b5520 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.472072] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 710.472292] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 710.473161] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9991a8a-b74e-45d3-a743-b3d9a5349c15 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.485160] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 710.485160] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bb8c1c-b828-e6d6-4508-839ddf199371" [ 710.485160] env[61728]: _type = "Task" [ 710.485160] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.494852] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bb8c1c-b828-e6d6-4508-839ddf199371, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.559499] env[61728]: DEBUG oslo_concurrency.lockutils [req-714416d9-c3fb-4026-b2e9-53b8dff1035b req-7d3305c3-fee2-47cb-a4e8-9b91fc1b9eb6 service nova] Releasing lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.559752] env[61728]: DEBUG nova.compute.manager [req-714416d9-c3fb-4026-b2e9-53b8dff1035b req-7d3305c3-fee2-47cb-a4e8-9b91fc1b9eb6 service nova] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Received event network-vif-deleted-ec661d56-ab1e-4593-8990-bf9e698c8880 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 710.660880] env[61728]: DEBUG nova.network.neutron [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Updating instance_info_cache with network_info: [{"id": "8fcb3f51-74e2-4548-8b86-bc5bc9156f40", "address": "fa:16:3e:e7:9f:e8", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.37", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fcb3f51-74", "ovs_interfaceid": "8fcb3f51-74e2-4548-8b86-bc5bc9156f40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.686305] env[61728]: DEBUG nova.compute.manager [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 710.802325] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463963, 'name': Rename_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.806175] env[61728]: DEBUG nova.network.neutron [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Successfully created port: 551e9af7-6eba-46a7-89a6-27f48212864d {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 710.859664] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463964, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.926792] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3721a741-4471-41a3-9939-b3f321623e2a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.935517] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8923872a-9fef-450f-aa58-d2fb225c591f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.974125] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1faa1c-d40d-45eb-a589-c2b8a395f46c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.981666] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf52f26-6069-4ecc-af8d-a3cec6a66651 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.995640] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bb8c1c-b828-e6d6-4508-839ddf199371, 'name': SearchDatastore_Task, 'duration_secs': 0.015024} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.004230] env[61728]: DEBUG nova.compute.provider_tree [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.009020] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-705a970a-e9d2-44b8-98c4-13f1074e3591 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.012047] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 711.012047] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5232a2c7-d3d3-cdf8-3b67-63eb93d12c01" [ 711.012047] env[61728]: _type = "Task" [ 711.012047] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.020947] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5232a2c7-d3d3-cdf8-3b67-63eb93d12c01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.166358] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Releasing lock "refresh_cache-bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.166864] env[61728]: DEBUG nova.compute.manager [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Instance network_info: |[{"id": "8fcb3f51-74e2-4548-8b86-bc5bc9156f40", "address": "fa:16:3e:e7:9f:e8", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.37", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fcb3f51-74", "ovs_interfaceid": "8fcb3f51-74e2-4548-8b86-bc5bc9156f40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 711.167527] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:9f:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '457c42cd-4ddb-4374-923e-d419b7f6eaff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8fcb3f51-74e2-4548-8b86-bc5bc9156f40', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 711.184524] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Creating folder: Project (68843e3f2b9a4e2fb9534db4fe5c2ec7). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 711.184949] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8bb3c0ce-1689-43fc-b755-be6a07f8b383 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.202843] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Created folder: Project (68843e3f2b9a4e2fb9534db4fe5c2ec7) in parent group-v121913. [ 711.203208] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Creating folder: Instances. Parent ref: group-v122060. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 711.203550] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-452339cd-864f-4f35-a00f-82538a1f4254 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.219244] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Created folder: Instances in parent group-v122060. [ 711.219512] env[61728]: DEBUG oslo.service.loopingcall [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.219751] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 711.219985] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-605871e4-b913-4510-8a2b-c1112b8377fb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.245477] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 711.245477] env[61728]: value = "task-463967" [ 711.245477] env[61728]: _type = "Task" [ 711.245477] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.255161] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463967, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.299479] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463963, 'name': Rename_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.358395] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463964, 'name': ReconfigVM_Task, 'duration_secs': 0.58711} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.358756] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Reconfigured VM instance instance-00000004 to attach disk [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4/771341ed-8b8e-470c-9686-82650f5271b4.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 711.359682] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d16db8a-fc42-4b59-a9e7-bdec80cbcc70 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.367477] env[61728]: DEBUG nova.compute.manager [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Received event network-vif-plugged-8fcb3f51-74e2-4548-8b86-bc5bc9156f40 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 711.367981] env[61728]: DEBUG oslo_concurrency.lockutils [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] Acquiring lock "bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.368151] env[61728]: DEBUG oslo_concurrency.lockutils [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] Lock "bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.368492] env[61728]: DEBUG oslo_concurrency.lockutils [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] Lock "bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.368492] env[61728]: DEBUG nova.compute.manager [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] No waiting events found dispatching network-vif-plugged-8fcb3f51-74e2-4548-8b86-bc5bc9156f40 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 711.368658] env[61728]: WARNING nova.compute.manager [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Received unexpected event network-vif-plugged-8fcb3f51-74e2-4548-8b86-bc5bc9156f40 for instance with vm_state building and task_state spawning. [ 711.368851] env[61728]: DEBUG nova.compute.manager [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Received event network-changed-8fcb3f51-74e2-4548-8b86-bc5bc9156f40 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 711.369059] env[61728]: DEBUG nova.compute.manager [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Refreshing instance network info cache due to event network-changed-8fcb3f51-74e2-4548-8b86-bc5bc9156f40. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 711.369290] env[61728]: DEBUG oslo_concurrency.lockutils [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] Acquiring lock "refresh_cache-bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.369633] env[61728]: DEBUG oslo_concurrency.lockutils [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] Acquired lock "refresh_cache-bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.369845] env[61728]: DEBUG nova.network.neutron [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Refreshing network info cache for port 8fcb3f51-74e2-4548-8b86-bc5bc9156f40 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 711.372838] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 711.372838] env[61728]: value = "task-463968" [ 711.372838] env[61728]: _type = "Task" [ 711.372838] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.383658] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463968, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.510240] env[61728]: DEBUG nova.scheduler.client.report [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.529322] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5232a2c7-d3d3-cdf8-3b67-63eb93d12c01, 'name': SearchDatastore_Task, 'duration_secs': 0.01464} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.529322] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.529523] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 223b1f93-d33e-4566-8c8e-75234a491814/223b1f93-d33e-4566-8c8e-75234a491814.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 711.531536] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d242e740-f0d6-4719-9cbd-f0df06fa8b8a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.543924] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 711.543924] env[61728]: value = "task-463969" [ 711.543924] env[61728]: _type = "Task" [ 711.543924] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.555162] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463969, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.701693] env[61728]: DEBUG nova.compute.manager [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 711.735401] env[61728]: DEBUG nova.virt.hardware [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 711.735401] env[61728]: DEBUG nova.virt.hardware [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 711.735401] env[61728]: DEBUG nova.virt.hardware [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 711.735401] env[61728]: DEBUG nova.virt.hardware [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 711.735401] env[61728]: DEBUG nova.virt.hardware [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 711.735401] env[61728]: DEBUG nova.virt.hardware [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 711.735401] env[61728]: DEBUG nova.virt.hardware [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 711.735704] env[61728]: DEBUG nova.virt.hardware [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 711.735704] env[61728]: DEBUG nova.virt.hardware [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 711.735901] env[61728]: DEBUG nova.virt.hardware [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 711.736169] env[61728]: DEBUG nova.virt.hardware [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 711.737268] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28574b4-1ce4-4b1b-8e85-6c835e4fecd4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.740117] env[61728]: DEBUG nova.network.neutron [-] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.752412] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a627167a-d5f0-40e2-8f8b-0feaa8686494 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.764982] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463967, 'name': CreateVM_Task, 'duration_secs': 0.45341} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.773172] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 711.775061] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.775061] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.775061] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 711.775317] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bc8f980-5d30-4ee6-98f2-a4c04b22aeeb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.781609] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 711.781609] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52afbbb6-8692-1a78-6ce7-14d8cba1ecd5" [ 711.781609] env[61728]: _type = "Task" [ 711.781609] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.792925] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52afbbb6-8692-1a78-6ce7-14d8cba1ecd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.804272] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463963, 'name': Rename_Task, 'duration_secs': 1.159942} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.804272] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 711.804601] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0a28c36-76be-4c08-871d-cab61a90d6cf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.813522] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 711.813522] env[61728]: value = "task-463970" [ 711.813522] env[61728]: _type = "Task" [ 711.813522] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.824704] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463970, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.895278] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463968, 'name': Rename_Task, 'duration_secs': 0.182975} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.895600] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 711.896299] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e3609d6a-254c-4d64-844e-b79399621f0e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.907108] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 711.907108] env[61728]: value = "task-463971" [ 711.907108] env[61728]: _type = "Task" [ 711.907108] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.921098] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463971, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.019592] env[61728]: DEBUG oslo_concurrency.lockutils [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.342s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.022637] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.262s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.025263] env[61728]: INFO nova.compute.claims [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.052965] env[61728]: INFO nova.scheduler.client.report [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Deleted allocations for instance 620bb472-c36e-4c56-acdf-42e02614856b [ 712.062372] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463969, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.165869] env[61728]: DEBUG nova.network.neutron [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Updated VIF entry in instance network info cache for port 8fcb3f51-74e2-4548-8b86-bc5bc9156f40. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 712.166429] env[61728]: DEBUG nova.network.neutron [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Updating instance_info_cache with network_info: [{"id": "8fcb3f51-74e2-4548-8b86-bc5bc9156f40", "address": "fa:16:3e:e7:9f:e8", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.37", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fcb3f51-74", "ovs_interfaceid": "8fcb3f51-74e2-4548-8b86-bc5bc9156f40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.243153] env[61728]: INFO nova.compute.manager [-] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Took 1.86 seconds to deallocate network for instance. [ 712.294178] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52afbbb6-8692-1a78-6ce7-14d8cba1ecd5, 'name': SearchDatastore_Task, 'duration_secs': 0.010403} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.295097] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.295097] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 712.295097] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.295277] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.295397] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 712.295608] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a25d731-1cae-4adf-8a6d-dc35ad8fd93a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.308022] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 712.308022] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 712.308022] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ae51ed2-95c6-42f0-8fa0-3f4d3421585d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.315139] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 712.315139] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52173f72-d02b-3169-07e5-53bef9792265" [ 712.315139] env[61728]: _type = "Task" [ 712.315139] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.332619] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52173f72-d02b-3169-07e5-53bef9792265, 'name': SearchDatastore_Task, 'duration_secs': 0.010267} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.333383] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463970, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.336017] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e7ddd42-f84f-4e0d-8573-972ebb59b209 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.342347] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 712.342347] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ed7722-5f1b-f074-8325-f5e584582709" [ 712.342347] env[61728]: _type = "Task" [ 712.342347] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.352968] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ed7722-5f1b-f074-8325-f5e584582709, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.422177] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463971, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.524534] env[61728]: DEBUG nova.network.neutron [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Successfully updated port: 551e9af7-6eba-46a7-89a6-27f48212864d {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 712.558049] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463969, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.680968} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.558472] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 223b1f93-d33e-4566-8c8e-75234a491814/223b1f93-d33e-4566-8c8e-75234a491814.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 712.558816] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 712.561355] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18867daf-b96a-43a8-9b10-c0ac0d3a08d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.564089] env[61728]: DEBUG oslo_concurrency.lockutils [None req-139131c3-8ffe-4d48-92ed-a42fefd7f848 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "620bb472-c36e-4c56-acdf-42e02614856b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.286s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.565055] env[61728]: DEBUG oslo_concurrency.lockutils [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] Acquired lock "620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.566432] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b9bd1f-fa97-4d5a-b892-f16162f4fb94 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.571687] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 712.571687] env[61728]: value = "task-463972" [ 712.571687] env[61728]: _type = "Task" [ 712.571687] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.575985] env[61728]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 712.576237] env[61728]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=61728) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 712.577134] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c8309a4-50bf-4132-86ca-140159715ba6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.583593] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463972, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.593927] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b725c8-c2e6-4f57-8212-bf5ce3da6f53 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.633848] env[61728]: ERROR root [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-121958' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-121958' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-121958' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-121958'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-121958' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-121958' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-121958'}\n"]: nova.exception.InstanceNotFound: Instance 620bb472-c36e-4c56-acdf-42e02614856b could not be found. [ 712.634097] env[61728]: DEBUG oslo_concurrency.lockutils [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] Releasing lock "620bb472-c36e-4c56-acdf-42e02614856b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.634329] env[61728]: DEBUG nova.compute.manager [req-a514fc60-29a0-493b-b42b-a12658a7dcda req-85c897d3-d021-4fb6-886a-2b35ae34ea61 service nova] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Detach interface failed, port_id=f605e2a3-ef9e-4642-8d18-e43436149349, reason: Instance 620bb472-c36e-4c56-acdf-42e02614856b could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 712.670305] env[61728]: DEBUG oslo_concurrency.lockutils [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] Releasing lock "refresh_cache-bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.670609] env[61728]: DEBUG nova.compute.manager [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Received event network-vif-deleted-8ee9bdfc-796a-498b-845f-02218a86801a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 712.670825] env[61728]: INFO nova.compute.manager [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Neutron deleted interface 8ee9bdfc-796a-498b-845f-02218a86801a; detaching it from the instance and deleting it from the info cache [ 712.671053] env[61728]: DEBUG nova.network.neutron [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.750191] env[61728]: DEBUG oslo_concurrency.lockutils [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.827593] env[61728]: DEBUG oslo_vmware.api [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463970, 'name': PowerOnVM_Task, 'duration_secs': 0.64901} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.827829] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 712.827960] env[61728]: INFO nova.compute.manager [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Took 9.59 seconds to spawn the instance on the hypervisor. [ 712.828159] env[61728]: DEBUG nova.compute.manager [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 712.828996] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ce747b-cf29-4c73-8c79-4b5503af2aa0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.855025] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ed7722-5f1b-f074-8325-f5e584582709, 'name': SearchDatastore_Task, 'duration_secs': 0.034236} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.855025] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.855265] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5/bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 712.855640] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cdd9b138-42f8-4929-8af0-27c607167b44 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.864947] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 712.864947] env[61728]: value = "task-463973" [ 712.864947] env[61728]: _type = "Task" [ 712.864947] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.877590] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463973, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.918790] env[61728]: DEBUG oslo_vmware.api [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463971, 'name': PowerOnVM_Task, 'duration_secs': 0.920763} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.919103] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 712.919329] env[61728]: DEBUG nova.compute.manager [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 712.920501] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-669c88de-a9ac-49f9-8d26-b8fbc3dbecb5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.028168] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "refresh_cache-6df7b619-8cc4-4dd9-8596-22dc83234a8e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.028342] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquired lock "refresh_cache-6df7b619-8cc4-4dd9-8596-22dc83234a8e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.028511] env[61728]: DEBUG nova.network.neutron [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 713.083468] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463972, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.206353} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.086260] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 713.087455] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18201fc-73d6-45c2-9d69-67127b9ce6f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.125469] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] 223b1f93-d33e-4566-8c8e-75234a491814/223b1f93-d33e-4566-8c8e-75234a491814.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 713.128673] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94153b6c-9962-45b8-9b9e-524e8d374959 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.152461] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 713.152461] env[61728]: value = "task-463974" [ 713.152461] env[61728]: _type = "Task" [ 713.152461] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.166393] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463974, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.174239] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b873371c-bc61-4f31-8396-1ca8b39a8586 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.189948] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-922a23ae-068f-433b-a91c-32b51a91e57f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.229930] env[61728]: DEBUG nova.compute.manager [req-1db7e6cd-77b4-4abb-b46d-4b9707c4bd85 req-abe19763-a84f-484f-9c26-70696b209903 service nova] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Detach interface failed, port_id=8ee9bdfc-796a-498b-845f-02218a86801a, reason: Instance 8c53c9c9-cdc1-407b-8827-8409cf137235 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 713.352407] env[61728]: INFO nova.compute.manager [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Took 53.43 seconds to build instance. [ 713.383330] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463973, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.442098] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.578819] env[61728]: DEBUG nova.network.neutron [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.646459] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c08841f7-8efa-4642-a5c3-ddd00c497222 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.661908] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2de8e65-b885-4ea4-a79c-049f463169e4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.668109] env[61728]: DEBUG nova.compute.manager [req-a5a97765-de22-4453-9472-e19fae8871c6 req-b5ed3429-3ac2-4d8b-b13c-aae44a5ab8c8 service nova] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Received event network-vif-plugged-551e9af7-6eba-46a7-89a6-27f48212864d {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 713.668109] env[61728]: DEBUG oslo_concurrency.lockutils [req-a5a97765-de22-4453-9472-e19fae8871c6 req-b5ed3429-3ac2-4d8b-b13c-aae44a5ab8c8 service nova] Acquiring lock "6df7b619-8cc4-4dd9-8596-22dc83234a8e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.668109] env[61728]: DEBUG oslo_concurrency.lockutils [req-a5a97765-de22-4453-9472-e19fae8871c6 req-b5ed3429-3ac2-4d8b-b13c-aae44a5ab8c8 service nova] Lock "6df7b619-8cc4-4dd9-8596-22dc83234a8e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.668377] env[61728]: DEBUG oslo_concurrency.lockutils [req-a5a97765-de22-4453-9472-e19fae8871c6 req-b5ed3429-3ac2-4d8b-b13c-aae44a5ab8c8 service nova] Lock "6df7b619-8cc4-4dd9-8596-22dc83234a8e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.668561] env[61728]: DEBUG nova.compute.manager [req-a5a97765-de22-4453-9472-e19fae8871c6 req-b5ed3429-3ac2-4d8b-b13c-aae44a5ab8c8 service nova] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] No waiting events found dispatching network-vif-plugged-551e9af7-6eba-46a7-89a6-27f48212864d {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 713.668779] env[61728]: WARNING nova.compute.manager [req-a5a97765-de22-4453-9472-e19fae8871c6 req-b5ed3429-3ac2-4d8b-b13c-aae44a5ab8c8 service nova] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Received unexpected event network-vif-plugged-551e9af7-6eba-46a7-89a6-27f48212864d for instance with vm_state building and task_state spawning. [ 713.668982] env[61728]: DEBUG nova.compute.manager [req-a5a97765-de22-4453-9472-e19fae8871c6 req-b5ed3429-3ac2-4d8b-b13c-aae44a5ab8c8 service nova] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Received event network-changed-551e9af7-6eba-46a7-89a6-27f48212864d {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 713.669207] env[61728]: DEBUG nova.compute.manager [req-a5a97765-de22-4453-9472-e19fae8871c6 req-b5ed3429-3ac2-4d8b-b13c-aae44a5ab8c8 service nova] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Refreshing instance network info cache due to event network-changed-551e9af7-6eba-46a7-89a6-27f48212864d. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 713.670143] env[61728]: DEBUG oslo_concurrency.lockutils [req-a5a97765-de22-4453-9472-e19fae8871c6 req-b5ed3429-3ac2-4d8b-b13c-aae44a5ab8c8 service nova] Acquiring lock "refresh_cache-6df7b619-8cc4-4dd9-8596-22dc83234a8e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.676329] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463974, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.704975] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f267501-2858-451a-bb41-fc75dfc4c962 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.713735] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017f7d58-1621-43b2-9ccf-9b9b88fe236c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.727915] env[61728]: DEBUG nova.compute.provider_tree [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.741567] env[61728]: DEBUG nova.network.neutron [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Updating instance_info_cache with network_info: [{"id": "551e9af7-6eba-46a7-89a6-27f48212864d", "address": "fa:16:3e:17:5d:8a", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.224", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap551e9af7-6e", "ovs_interfaceid": "551e9af7-6eba-46a7-89a6-27f48212864d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.859660] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1eead96a-70c3-4c98-bcf4-599649acf2c3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "c14034b5-330b-4fb1-af31-d530e3231992" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.440s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.882796] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463973, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.165954] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463974, 'name': ReconfigVM_Task, 'duration_secs': 1.00328} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.166920] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Reconfigured VM instance instance-00000030 to attach disk [datastore1] 223b1f93-d33e-4566-8c8e-75234a491814/223b1f93-d33e-4566-8c8e-75234a491814.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 714.167711] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ac2ed514-2780-4bdc-9919-1ed01f4c18a7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.178229] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 714.178229] env[61728]: value = "task-463975" [ 714.178229] env[61728]: _type = "Task" [ 714.178229] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.187844] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463975, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.233128] env[61728]: DEBUG nova.scheduler.client.report [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.244598] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Releasing lock "refresh_cache-6df7b619-8cc4-4dd9-8596-22dc83234a8e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.245548] env[61728]: DEBUG nova.compute.manager [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Instance network_info: |[{"id": "551e9af7-6eba-46a7-89a6-27f48212864d", "address": "fa:16:3e:17:5d:8a", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.224", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap551e9af7-6e", "ovs_interfaceid": "551e9af7-6eba-46a7-89a6-27f48212864d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 714.245548] env[61728]: DEBUG oslo_concurrency.lockutils [req-a5a97765-de22-4453-9472-e19fae8871c6 req-b5ed3429-3ac2-4d8b-b13c-aae44a5ab8c8 service nova] Acquired lock "refresh_cache-6df7b619-8cc4-4dd9-8596-22dc83234a8e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.245548] env[61728]: DEBUG nova.network.neutron [req-a5a97765-de22-4453-9472-e19fae8871c6 req-b5ed3429-3ac2-4d8b-b13c-aae44a5ab8c8 service nova] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Refreshing network info cache for port 551e9af7-6eba-46a7-89a6-27f48212864d {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 714.246592] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:5d:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '457c42cd-4ddb-4374-923e-d419b7f6eaff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '551e9af7-6eba-46a7-89a6-27f48212864d', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 714.256020] env[61728]: DEBUG oslo.service.loopingcall [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 714.259259] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 714.259259] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b0649051-90fb-4ce3-8724-502084050bc3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.283071] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 714.283071] env[61728]: value = "task-463976" [ 714.283071] env[61728]: _type = "Task" [ 714.283071] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.291946] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463976, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.363471] env[61728]: DEBUG nova.compute.manager [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 714.384732] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463973, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.172759} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.384732] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5/bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 714.384894] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 714.385177] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad7eb4fd-bae9-4543-bb58-3b459f7e2fa1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.394439] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 714.394439] env[61728]: value = "task-463977" [ 714.394439] env[61728]: _type = "Task" [ 714.394439] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.406918] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463977, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.691026] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463975, 'name': Rename_Task, 'duration_secs': 0.168245} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.691296] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 714.691980] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5fea923a-5ad0-4a95-80a0-61432b0a132b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.699275] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 714.699275] env[61728]: value = "task-463978" [ 714.699275] env[61728]: _type = "Task" [ 714.699275] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.708125] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463978, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.738047] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.715s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.739043] env[61728]: DEBUG nova.compute.manager [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 714.741463] env[61728]: DEBUG oslo_concurrency.lockutils [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.627s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.741637] env[61728]: DEBUG nova.objects.instance [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lazy-loading 'resources' on Instance uuid 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 714.795201] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463976, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.885104] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40a33c89-1591-46d3-99f6-3fc758fc1d65 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.892422] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-370a7582-2f49-4a63-8d7b-372d120f3bf3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Suspending the VM {{(pid=61728) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 714.896742] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-0c319fd7-890d-492a-8cc4-6ed50e379183 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.899186] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.916799] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463977, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086125} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.918289] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 714.918680] env[61728]: DEBUG oslo_vmware.api [None req-370a7582-2f49-4a63-8d7b-372d120f3bf3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 714.918680] env[61728]: value = "task-463979" [ 714.918680] env[61728]: _type = "Task" [ 714.918680] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.919476] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3596159f-6a68-4e36-9f78-0ab0f425b80b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.936877] env[61728]: DEBUG oslo_vmware.api [None req-370a7582-2f49-4a63-8d7b-372d120f3bf3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463979, 'name': SuspendVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.957670] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5/bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 714.960534] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d4e84f5-7d2e-43f3-ac30-0b74ef3668b2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.985548] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 714.985548] env[61728]: value = "task-463980" [ 714.985548] env[61728]: _type = "Task" [ 714.985548] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.993324] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463980, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.106566] env[61728]: DEBUG nova.network.neutron [req-a5a97765-de22-4453-9472-e19fae8871c6 req-b5ed3429-3ac2-4d8b-b13c-aae44a5ab8c8 service nova] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Updated VIF entry in instance network info cache for port 551e9af7-6eba-46a7-89a6-27f48212864d. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 715.106949] env[61728]: DEBUG nova.network.neutron [req-a5a97765-de22-4453-9472-e19fae8871c6 req-b5ed3429-3ac2-4d8b-b13c-aae44a5ab8c8 service nova] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Updating instance_info_cache with network_info: [{"id": "551e9af7-6eba-46a7-89a6-27f48212864d", "address": "fa:16:3e:17:5d:8a", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.224", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap551e9af7-6e", "ovs_interfaceid": "551e9af7-6eba-46a7-89a6-27f48212864d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.214456] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463978, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.248257] env[61728]: DEBUG nova.compute.utils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.250157] env[61728]: DEBUG nova.compute.manager [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 715.250340] env[61728]: DEBUG nova.network.neutron [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 715.303046] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463976, 'name': CreateVM_Task, 'duration_secs': 0.808536} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.303555] env[61728]: DEBUG nova.policy [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd53ec83e72e44563946aa13db7e7ca02', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '29dfac4a253f43be811df957e0ee5f96', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 715.306198] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 715.307124] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.307338] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.307768] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 715.308421] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55a92206-fff6-421c-884b-5ad0b5ff6696 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.318465] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 715.318465] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5230af18-0862-36fc-bdb4-8e3d5894e83d" [ 715.318465] env[61728]: _type = "Task" [ 715.318465] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.329138] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5230af18-0862-36fc-bdb4-8e3d5894e83d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.433292] env[61728]: DEBUG oslo_vmware.api [None req-370a7582-2f49-4a63-8d7b-372d120f3bf3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463979, 'name': SuspendVM_Task} progress is 58%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.493762] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463980, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.613767] env[61728]: DEBUG oslo_concurrency.lockutils [req-a5a97765-de22-4453-9472-e19fae8871c6 req-b5ed3429-3ac2-4d8b-b13c-aae44a5ab8c8 service nova] Releasing lock "refresh_cache-6df7b619-8cc4-4dd9-8596-22dc83234a8e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.653470] env[61728]: DEBUG nova.network.neutron [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Successfully created port: 372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.715348] env[61728]: DEBUG oslo_vmware.api [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463978, 'name': PowerOnVM_Task, 'duration_secs': 0.880387} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.715725] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 715.715899] env[61728]: INFO nova.compute.manager [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Took 9.62 seconds to spawn the instance on the hypervisor. [ 715.716103] env[61728]: DEBUG nova.compute.manager [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 715.717834] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8ab245-85c7-45bf-bd71-b0034dff242d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.756904] env[61728]: DEBUG nova.compute.manager [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 715.806615] env[61728]: DEBUG oslo_concurrency.lockutils [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "b833f691-e76e-4a2e-94a4-7594fadc3ba8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.807538] env[61728]: DEBUG oslo_concurrency.lockutils [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "b833f691-e76e-4a2e-94a4-7594fadc3ba8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.807538] env[61728]: DEBUG oslo_concurrency.lockutils [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "b833f691-e76e-4a2e-94a4-7594fadc3ba8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.807538] env[61728]: DEBUG oslo_concurrency.lockutils [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "b833f691-e76e-4a2e-94a4-7594fadc3ba8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.807538] env[61728]: DEBUG oslo_concurrency.lockutils [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "b833f691-e76e-4a2e-94a4-7594fadc3ba8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.814915] env[61728]: INFO nova.compute.manager [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Terminating instance [ 715.819411] env[61728]: DEBUG nova.compute.manager [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 715.819666] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 715.821100] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86125ba1-f61e-4ce8-ad06-7e9cf7f95261 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.838552] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5230af18-0862-36fc-bdb4-8e3d5894e83d, 'name': SearchDatastore_Task, 'duration_secs': 0.013413} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.841287] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.841598] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 715.841885] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.842092] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.842370] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 715.842621] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 715.842881] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99b01581-77db-459e-9de6-d43b20cc9ea0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.845617] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4656a2ab-a167-4927-8272-361ae35a2d46 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.854694] env[61728]: DEBUG oslo_vmware.api [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 715.854694] env[61728]: value = "task-463981" [ 715.854694] env[61728]: _type = "Task" [ 715.854694] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.859688] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 715.859907] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 715.861019] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-744dd13c-5759-4f86-a29b-4a39303b3427 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.866162] env[61728]: DEBUG oslo_vmware.api [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463981, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.871127] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 715.871127] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526071ff-3dc1-0b75-0aa7-95f1ba750181" [ 715.871127] env[61728]: _type = "Task" [ 715.871127] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.881791] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526071ff-3dc1-0b75-0aa7-95f1ba750181, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.935215] env[61728]: DEBUG oslo_vmware.api [None req-370a7582-2f49-4a63-8d7b-372d120f3bf3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463979, 'name': SuspendVM_Task, 'duration_secs': 0.69694} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.939631] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-370a7582-2f49-4a63-8d7b-372d120f3bf3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Suspended the VM {{(pid=61728) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 715.939631] env[61728]: DEBUG nova.compute.manager [None req-370a7582-2f49-4a63-8d7b-372d120f3bf3 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 715.940850] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0292cee-d209-49f7-8121-5ad1bce71268 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.999369] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463980, 'name': ReconfigVM_Task, 'duration_secs': 0.548055} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.000726] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Reconfigured VM instance instance-00000031 to attach disk [datastore1] bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5/bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 716.001853] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-285b7155-ea1f-4881-9585-78f509fd41a5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.004749] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-07317bec-00fb-4e09-8a19-25d2a8338529 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.012961] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b55064be-569d-4c65-a16b-d03a978ee670 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.018402] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 716.018402] env[61728]: value = "task-463982" [ 716.018402] env[61728]: _type = "Task" [ 716.018402] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.051377] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4682974-bb02-4ced-9a5a-330fc6b5af6f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.057776] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463982, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.063652] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab49298-30d1-4227-ac32-e1a0fa660195 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.081924] env[61728]: DEBUG nova.compute.provider_tree [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.244260] env[61728]: INFO nova.compute.manager [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Took 38.05 seconds to build instance. [ 716.246683] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "f28348d2-c062-497a-b374-521df51054ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.246683] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "f28348d2-c062-497a-b374-521df51054ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.365834] env[61728]: DEBUG oslo_vmware.api [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463981, 'name': PowerOffVM_Task, 'duration_secs': 0.268354} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.366146] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 716.366905] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 716.366905] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-94404516-b07a-4501-9744-f77f2fbc907f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.385045] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526071ff-3dc1-0b75-0aa7-95f1ba750181, 'name': SearchDatastore_Task, 'duration_secs': 0.030704} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.388856] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d31412c-311e-4259-a5c5-ab7f318de31a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.390969] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 716.390969] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bcfa42-0425-8b59-a830-6afac38a34f7" [ 716.390969] env[61728]: _type = "Task" [ 716.390969] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.399789] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bcfa42-0425-8b59-a830-6afac38a34f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.445851] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 716.445851] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 716.446292] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleting the datastore file [datastore1] b833f691-e76e-4a2e-94a4-7594fadc3ba8 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 716.446334] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9a0444a8-782c-4aae-ac35-5cab884b5679 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.457091] env[61728]: DEBUG oslo_vmware.api [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 716.457091] env[61728]: value = "task-463984" [ 716.457091] env[61728]: _type = "Task" [ 716.457091] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.466484] env[61728]: DEBUG oslo_vmware.api [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463984, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.530508] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463982, 'name': Rename_Task, 'duration_secs': 0.383511} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.530775] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 716.531075] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d7299adf-8458-487b-b6ab-d63e2f57ac77 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.539825] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 716.539825] env[61728]: value = "task-463985" [ 716.539825] env[61728]: _type = "Task" [ 716.539825] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.557122] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463985, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.588094] env[61728]: DEBUG nova.scheduler.client.report [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.748276] env[61728]: DEBUG oslo_concurrency.lockutils [None req-900edbb4-74e3-4e19-b34e-8bb214ff3fb5 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "223b1f93-d33e-4566-8c8e-75234a491814" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.937s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.766204] env[61728]: DEBUG nova.compute.manager [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 716.796573] env[61728]: DEBUG nova.virt.hardware [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.796923] env[61728]: DEBUG nova.virt.hardware [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.797155] env[61728]: DEBUG nova.virt.hardware [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.797383] env[61728]: DEBUG nova.virt.hardware [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.797556] env[61728]: DEBUG nova.virt.hardware [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.797731] env[61728]: DEBUG nova.virt.hardware [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.798059] env[61728]: DEBUG nova.virt.hardware [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.798305] env[61728]: DEBUG nova.virt.hardware [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.798802] env[61728]: DEBUG nova.virt.hardware [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.799059] env[61728]: DEBUG nova.virt.hardware [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.799294] env[61728]: DEBUG nova.virt.hardware [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.800643] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5274819-f3ca-4ddf-a506-0fc3febafa8f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.812121] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57441e8-fbcd-4c0e-82ca-3306020cdd98 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.908060] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bcfa42-0425-8b59-a830-6afac38a34f7, 'name': SearchDatastore_Task, 'duration_secs': 0.034823} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.908388] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.908717] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 6df7b619-8cc4-4dd9-8596-22dc83234a8e/6df7b619-8cc4-4dd9-8596-22dc83234a8e.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 716.909653] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c3c68282-fff4-4aab-b0af-2ac12a466516 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.919685] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 716.919685] env[61728]: value = "task-463986" [ 716.919685] env[61728]: _type = "Task" [ 716.919685] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.929837] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463986, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.970814] env[61728]: DEBUG oslo_vmware.api [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-463984, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.358663} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.971382] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 716.971696] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 716.971932] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 716.972171] env[61728]: INFO nova.compute.manager [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 716.972439] env[61728]: DEBUG oslo.service.loopingcall [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.972680] env[61728]: DEBUG nova.compute.manager [-] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.972846] env[61728]: DEBUG nova.network.neutron [-] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 717.051079] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463985, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.065311] env[61728]: DEBUG nova.compute.manager [req-35a55cd6-fadb-4f87-9436-d13f271ea66d req-e311ef97-8f18-4f2a-9aa1-c330ea38f61b service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Received event network-changed-401726ae-678d-4918-b8c1-441b3802ddc8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 717.065520] env[61728]: DEBUG nova.compute.manager [req-35a55cd6-fadb-4f87-9436-d13f271ea66d req-e311ef97-8f18-4f2a-9aa1-c330ea38f61b service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Refreshing instance network info cache due to event network-changed-401726ae-678d-4918-b8c1-441b3802ddc8. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 717.065759] env[61728]: DEBUG oslo_concurrency.lockutils [req-35a55cd6-fadb-4f87-9436-d13f271ea66d req-e311ef97-8f18-4f2a-9aa1-c330ea38f61b service nova] Acquiring lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.065928] env[61728]: DEBUG oslo_concurrency.lockutils [req-35a55cd6-fadb-4f87-9436-d13f271ea66d req-e311ef97-8f18-4f2a-9aa1-c330ea38f61b service nova] Acquired lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.066882] env[61728]: DEBUG nova.network.neutron [req-35a55cd6-fadb-4f87-9436-d13f271ea66d req-e311ef97-8f18-4f2a-9aa1-c330ea38f61b service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Refreshing network info cache for port 401726ae-678d-4918-b8c1-441b3802ddc8 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 717.092632] env[61728]: DEBUG oslo_concurrency.lockutils [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.351s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.094866] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.678s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.098942] env[61728]: INFO nova.compute.claims [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 717.127260] env[61728]: INFO nova.scheduler.client.report [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Deleted allocations for instance 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1 [ 717.251494] env[61728]: DEBUG nova.compute.manager [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 717.355849] env[61728]: DEBUG nova.compute.manager [req-9d8ebf59-9531-4a95-a13e-1179f034f04a req-a88d0233-bc6f-4991-8c05-2a5032ed7aec service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Received event network-vif-plugged-372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 717.356136] env[61728]: DEBUG oslo_concurrency.lockutils [req-9d8ebf59-9531-4a95-a13e-1179f034f04a req-a88d0233-bc6f-4991-8c05-2a5032ed7aec service nova] Acquiring lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.356441] env[61728]: DEBUG oslo_concurrency.lockutils [req-9d8ebf59-9531-4a95-a13e-1179f034f04a req-a88d0233-bc6f-4991-8c05-2a5032ed7aec service nova] Lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.356525] env[61728]: DEBUG oslo_concurrency.lockutils [req-9d8ebf59-9531-4a95-a13e-1179f034f04a req-a88d0233-bc6f-4991-8c05-2a5032ed7aec service nova] Lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.356668] env[61728]: DEBUG nova.compute.manager [req-9d8ebf59-9531-4a95-a13e-1179f034f04a req-a88d0233-bc6f-4991-8c05-2a5032ed7aec service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] No waiting events found dispatching network-vif-plugged-372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 717.356887] env[61728]: WARNING nova.compute.manager [req-9d8ebf59-9531-4a95-a13e-1179f034f04a req-a88d0233-bc6f-4991-8c05-2a5032ed7aec service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Received unexpected event network-vif-plugged-372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8 for instance with vm_state building and task_state spawning. [ 717.431497] env[61728]: DEBUG nova.network.neutron [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Successfully updated port: 372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 717.441172] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463986, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.551435] env[61728]: DEBUG oslo_vmware.api [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463985, 'name': PowerOnVM_Task, 'duration_secs': 0.873781} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.551933] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 717.552020] env[61728]: INFO nova.compute.manager [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Took 8.67 seconds to spawn the instance on the hypervisor. [ 717.552240] env[61728]: DEBUG nova.compute.manager [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 717.553120] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec9460c-4cad-4d26-b572-a85934f8fe1b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.636935] env[61728]: DEBUG oslo_concurrency.lockutils [None req-774626f2-3ce4-4836-b76f-8cef6a4d0312 tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.965s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.774146] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.781495] env[61728]: DEBUG nova.network.neutron [req-35a55cd6-fadb-4f87-9436-d13f271ea66d req-e311ef97-8f18-4f2a-9aa1-c330ea38f61b service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Updated VIF entry in instance network info cache for port 401726ae-678d-4918-b8c1-441b3802ddc8. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 717.781881] env[61728]: DEBUG nova.network.neutron [req-35a55cd6-fadb-4f87-9436-d13f271ea66d req-e311ef97-8f18-4f2a-9aa1-c330ea38f61b service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Updating instance_info_cache with network_info: [{"id": "401726ae-678d-4918-b8c1-441b3802ddc8", "address": "fa:16:3e:36:16:cf", "network": {"id": "5d1e643d-f8a4-4302-9920-bb661df4632d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1891316321-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fc16f1d1b144286b180cd7c8f7588fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap401726ae-67", "ovs_interfaceid": "401726ae-678d-4918-b8c1-441b3802ddc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.835319] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "223b1f93-d33e-4566-8c8e-75234a491814" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.835605] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "223b1f93-d33e-4566-8c8e-75234a491814" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.835853] env[61728]: INFO nova.compute.manager [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Rebooting instance [ 717.913380] env[61728]: DEBUG nova.network.neutron [-] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.932317] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463986, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.941733] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "refresh_cache-c84bfcd0-b145-4675-8b0a-5e8f94f65098" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.941733] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquired lock "refresh_cache-c84bfcd0-b145-4675-8b0a-5e8f94f65098" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.941733] env[61728]: DEBUG nova.network.neutron [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 718.076615] env[61728]: INFO nova.compute.manager [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Took 36.43 seconds to build instance. [ 718.285715] env[61728]: DEBUG oslo_concurrency.lockutils [req-35a55cd6-fadb-4f87-9436-d13f271ea66d req-e311ef97-8f18-4f2a-9aa1-c330ea38f61b service nova] Releasing lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.365397] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.365645] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquired lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.365778] env[61728]: DEBUG nova.network.neutron [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 718.415093] env[61728]: INFO nova.compute.manager [-] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Took 1.44 seconds to deallocate network for instance. [ 718.436995] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463986, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.492209] env[61728]: DEBUG nova.network.neutron [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.578940] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4c164782-5a44-40e1-bdd8-43866dfb0788 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.927s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.711131] env[61728]: DEBUG nova.network.neutron [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Updating instance_info_cache with network_info: [{"id": "372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8", "address": "fa:16:3e:68:e7:ba", "network": {"id": "22e73f4c-99d9-4e65-8a72-f6ea53d88eaf", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2136421023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "29dfac4a253f43be811df957e0ee5f96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap372c7b2a-4d", "ovs_interfaceid": "372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.774144] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233ac6ba-b3ac-47ab-9652-b706aa8653c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.784553] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc70a464-4a06-4548-a373-ce429452d383 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.824741] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1078a49c-b1a1-4667-9d37-57b16e0c9821 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.834627] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-070964d9-d15a-46cc-890a-1bc0210300de {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.852116] env[61728]: DEBUG nova.compute.provider_tree [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 718.919854] env[61728]: DEBUG nova.compute.manager [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 718.919854] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d1128e-6fc1-4c99-8bbb-2e82e1d840a8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.926582] env[61728]: DEBUG oslo_concurrency.lockutils [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.937317] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463986, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.709493} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.941171] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 6df7b619-8cc4-4dd9-8596-22dc83234a8e/6df7b619-8cc4-4dd9-8596-22dc83234a8e.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 718.941171] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 718.941171] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-157c0177-2890-4fa5-91a9-29c772db6ea6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.951853] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 718.951853] env[61728]: value = "task-463987" [ 718.951853] env[61728]: _type = "Task" [ 718.951853] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.958356] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463987, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.086881] env[61728]: DEBUG nova.compute.manager [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 719.215886] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Releasing lock "refresh_cache-c84bfcd0-b145-4675-8b0a-5e8f94f65098" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.216049] env[61728]: DEBUG nova.compute.manager [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Instance network_info: |[{"id": "372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8", "address": "fa:16:3e:68:e7:ba", "network": {"id": "22e73f4c-99d9-4e65-8a72-f6ea53d88eaf", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2136421023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "29dfac4a253f43be811df957e0ee5f96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap372c7b2a-4d", "ovs_interfaceid": "372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 719.216412] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:e7:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6685c85e-be1e-4b7b-a6cc-3e50e59b6567', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 719.225140] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Creating folder: Project (29dfac4a253f43be811df957e0ee5f96). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 719.225528] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-31b8cecd-80d7-41e9-a9a2-bcf17fc53f97 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.230957] env[61728]: DEBUG nova.network.neutron [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Updating instance_info_cache with network_info: [{"id": "401726ae-678d-4918-b8c1-441b3802ddc8", "address": "fa:16:3e:36:16:cf", "network": {"id": "5d1e643d-f8a4-4302-9920-bb661df4632d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1891316321-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fc16f1d1b144286b180cd7c8f7588fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap401726ae-67", "ovs_interfaceid": "401726ae-678d-4918-b8c1-441b3802ddc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.238567] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Created folder: Project (29dfac4a253f43be811df957e0ee5f96) in parent group-v121913. [ 719.238848] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Creating folder: Instances. Parent ref: group-v122064. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 719.239640] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bcbd6feb-5bd7-4034-a581-e4d55ae837f0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.250321] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Created folder: Instances in parent group-v122064. [ 719.250618] env[61728]: DEBUG oslo.service.loopingcall [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 719.251142] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 719.251413] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dca2f07a-a99f-4b78-a90e-7f42df273aa3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.275903] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 719.275903] env[61728]: value = "task-463990" [ 719.275903] env[61728]: _type = "Task" [ 719.275903] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.284517] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463990, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.375450] env[61728]: ERROR nova.scheduler.client.report [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [req-4706b8fd-0ee3-47ff-b938-5aa696fac14a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4706b8fd-0ee3-47ff-b938-5aa696fac14a"}]} [ 719.398213] env[61728]: DEBUG nova.scheduler.client.report [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 719.402587] env[61728]: DEBUG nova.compute.manager [req-20386b50-9e31-477c-a811-97cfc4b9af24 req-633ba44e-b3a0-4f74-a228-4cde3bb04f9a service nova] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Received event network-vif-deleted-a808f8cb-f136-434e-8bf7-5ee0b957e156 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 719.403192] env[61728]: DEBUG nova.compute.manager [req-20386b50-9e31-477c-a811-97cfc4b9af24 req-633ba44e-b3a0-4f74-a228-4cde3bb04f9a service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Received event network-changed-372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 719.403192] env[61728]: DEBUG nova.compute.manager [req-20386b50-9e31-477c-a811-97cfc4b9af24 req-633ba44e-b3a0-4f74-a228-4cde3bb04f9a service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Refreshing instance network info cache due to event network-changed-372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 719.403381] env[61728]: DEBUG oslo_concurrency.lockutils [req-20386b50-9e31-477c-a811-97cfc4b9af24 req-633ba44e-b3a0-4f74-a228-4cde3bb04f9a service nova] Acquiring lock "refresh_cache-c84bfcd0-b145-4675-8b0a-5e8f94f65098" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.403533] env[61728]: DEBUG oslo_concurrency.lockutils [req-20386b50-9e31-477c-a811-97cfc4b9af24 req-633ba44e-b3a0-4f74-a228-4cde3bb04f9a service nova] Acquired lock "refresh_cache-c84bfcd0-b145-4675-8b0a-5e8f94f65098" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.403760] env[61728]: DEBUG nova.network.neutron [req-20386b50-9e31-477c-a811-97cfc4b9af24 req-633ba44e-b3a0-4f74-a228-4cde3bb04f9a service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Refreshing network info cache for port 372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 719.413182] env[61728]: DEBUG nova.scheduler.client.report [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 719.413420] env[61728]: DEBUG nova.compute.provider_tree [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 719.427322] env[61728]: DEBUG nova.scheduler.client.report [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 719.443043] env[61728]: INFO nova.compute.manager [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] instance snapshotting [ 719.443043] env[61728]: WARNING nova.compute.manager [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 719.444433] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08c524b-8869-42f2-81f7-a82bab860836 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.449288] env[61728]: DEBUG nova.scheduler.client.report [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 719.470957] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a676a0-e3aa-47ff-a3c2-888dbdae4c40 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.479626] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463987, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07415} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.481623] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 719.487807] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd07a5e-1f7f-4bb7-9edd-d18c0246c15b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.511894] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 6df7b619-8cc4-4dd9-8596-22dc83234a8e/6df7b619-8cc4-4dd9-8596-22dc83234a8e.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 719.515248] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2fc0669-577d-4dbe-9c9a-f4f9f16e47b6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.537273] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 719.537273] env[61728]: value = "task-463991" [ 719.537273] env[61728]: _type = "Task" [ 719.537273] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.550042] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463991, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.607641] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.737256] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Releasing lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.738922] env[61728]: DEBUG nova.compute.manager [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 719.740317] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b5d7a9c-adf7-43d3-a3d7-71326d224806 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.793024] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-463990, 'name': CreateVM_Task, 'duration_secs': 0.370584} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.793024] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 719.793024] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.793024] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.793024] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 719.793024] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b6cfe71-2d2c-4b53-89ae-fc133b3c0fdf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.804039] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 719.804039] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5258aa95-65dc-47f3-5848-eeb73a5192d2" [ 719.804039] env[61728]: _type = "Task" [ 719.804039] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.815773] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5258aa95-65dc-47f3-5848-eeb73a5192d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.991675] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 719.994275] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-106886a5-01f4-4eec-aa8e-81dfa84f6493 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.004745] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 720.004745] env[61728]: value = "task-463992" [ 720.004745] env[61728]: _type = "Task" [ 720.004745] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.026150] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463992, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.050268] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463991, 'name': ReconfigVM_Task, 'duration_secs': 0.388843} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.050268] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 6df7b619-8cc4-4dd9-8596-22dc83234a8e/6df7b619-8cc4-4dd9-8596-22dc83234a8e.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 720.054391] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a6fa4ee0-f422-4252-8cb5-dd184f64d77c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.063235] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 720.063235] env[61728]: value = "task-463993" [ 720.063235] env[61728]: _type = "Task" [ 720.063235] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.075445] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463993, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.095894] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d50d3af-b2b6-447b-a6b5-2295b90fd917 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.104708] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f700286-b717-463d-b07c-bd5edee316a7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.146735] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e375d64c-e24c-488f-9925-f29b2d06fbee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.155920] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8919cff7-c863-4dea-a235-14fb6bcfb02c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.171318] env[61728]: DEBUG nova.compute.provider_tree [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 720.273809] env[61728]: DEBUG nova.network.neutron [req-20386b50-9e31-477c-a811-97cfc4b9af24 req-633ba44e-b3a0-4f74-a228-4cde3bb04f9a service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Updated VIF entry in instance network info cache for port 372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 720.274198] env[61728]: DEBUG nova.network.neutron [req-20386b50-9e31-477c-a811-97cfc4b9af24 req-633ba44e-b3a0-4f74-a228-4cde3bb04f9a service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Updating instance_info_cache with network_info: [{"id": "372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8", "address": "fa:16:3e:68:e7:ba", "network": {"id": "22e73f4c-99d9-4e65-8a72-f6ea53d88eaf", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2136421023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "29dfac4a253f43be811df957e0ee5f96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap372c7b2a-4d", "ovs_interfaceid": "372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.313807] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5258aa95-65dc-47f3-5848-eeb73a5192d2, 'name': SearchDatastore_Task, 'duration_secs': 0.016316} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.314120] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.314375] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 720.314617] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.314776] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.314968] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 720.315255] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-700eb9cf-89e9-487a-acff-20d65f9f475d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.327899] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 720.328126] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 720.328906] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf58dd79-6b4a-46bf-9a0d-3c2e680f1510 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.334933] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 720.334933] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a122bf-93a7-bceb-88c3-1de5a7d7f1f6" [ 720.334933] env[61728]: _type = "Task" [ 720.334933] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.345631] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a122bf-93a7-bceb-88c3-1de5a7d7f1f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.514842] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463992, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.573970] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463993, 'name': Rename_Task, 'duration_secs': 0.148426} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.574307] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 720.574611] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-451cf12e-1f12-495a-a8a8-b23671067b89 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.583077] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 720.583077] env[61728]: value = "task-463994" [ 720.583077] env[61728]: _type = "Task" [ 720.583077] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.592460] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463994, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.707932] env[61728]: DEBUG nova.scheduler.client.report [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 85 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 720.708496] env[61728]: DEBUG nova.compute.provider_tree [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 85 to 86 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 720.708713] env[61728]: DEBUG nova.compute.provider_tree [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 720.758396] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-840465ce-17ba-4881-9e09-c1d1f65b09c8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.767053] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Doing hard reboot of VM {{(pid=61728) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 720.767386] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-154bbb34-c2fb-48ea-a5f0-510afc9d32e2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.775611] env[61728]: DEBUG oslo_vmware.api [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 720.775611] env[61728]: value = "task-463995" [ 720.775611] env[61728]: _type = "Task" [ 720.775611] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.779358] env[61728]: DEBUG oslo_concurrency.lockutils [req-20386b50-9e31-477c-a811-97cfc4b9af24 req-633ba44e-b3a0-4f74-a228-4cde3bb04f9a service nova] Releasing lock "refresh_cache-c84bfcd0-b145-4675-8b0a-5e8f94f65098" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.787424] env[61728]: DEBUG oslo_vmware.api [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463995, 'name': ResetVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.831554] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 720.831554] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 720.846934] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a122bf-93a7-bceb-88c3-1de5a7d7f1f6, 'name': SearchDatastore_Task, 'duration_secs': 0.009865} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.846934] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29e2c375-fded-408b-b6dd-72585826e35a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.853201] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 720.853201] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526d21d2-fb1f-12c4-80d1-0da876ed015b" [ 720.853201] env[61728]: _type = "Task" [ 720.853201] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.862361] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526d21d2-fb1f-12c4-80d1-0da876ed015b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.021240] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463992, 'name': CreateSnapshot_Task, 'duration_secs': 0.630602} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.021514] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 721.023053] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b9b74d0-aabb-476a-839f-2c8744980d56 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.096370] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463994, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.218022] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.121s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.218022] env[61728]: DEBUG nova.compute.manager [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 721.219231] env[61728]: DEBUG oslo_concurrency.lockutils [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.972s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.219746] env[61728]: DEBUG nova.objects.instance [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lazy-loading 'resources' on Instance uuid 650d2875-68f7-4131-b9b7-9799ee08f1d6 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 721.288639] env[61728]: DEBUG oslo_vmware.api [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-463995, 'name': ResetVM_Task, 'duration_secs': 0.105361} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.289111] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Did hard reboot of VM {{(pid=61728) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 721.289424] env[61728]: DEBUG nova.compute.manager [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 721.290553] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03419f4-b292-4ffb-86e5-4f2950bd0d48 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.344761] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 721.345368] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Starting heal instance info cache {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 721.365900] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526d21d2-fb1f-12c4-80d1-0da876ed015b, 'name': SearchDatastore_Task, 'duration_secs': 0.012442} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.365900] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.366058] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] c84bfcd0-b145-4675-8b0a-5e8f94f65098/c84bfcd0-b145-4675-8b0a-5e8f94f65098.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 721.367032] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c933a8ff-c6cc-47a2-b1b4-5715645bb6db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.373682] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 721.373682] env[61728]: value = "task-463996" [ 721.373682] env[61728]: _type = "Task" [ 721.373682] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.383387] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-463996, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.548106] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 721.548454] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b929a1cf-cdcc-4c76-9085-083648e87e97 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.558797] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 721.558797] env[61728]: value = "task-463997" [ 721.558797] env[61728]: _type = "Task" [ 721.558797] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.571490] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463997, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.595403] env[61728]: DEBUG oslo_vmware.api [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-463994, 'name': PowerOnVM_Task, 'duration_secs': 0.798016} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.595785] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 721.596125] env[61728]: INFO nova.compute.manager [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Took 9.89 seconds to spawn the instance on the hypervisor. [ 721.596331] env[61728]: DEBUG nova.compute.manager [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 721.597422] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7198ad-51cc-4822-a1ab-f379291ee9ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.605769] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.606023] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.723304] env[61728]: DEBUG nova.compute.utils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.724958] env[61728]: DEBUG nova.compute.manager [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 721.725222] env[61728]: DEBUG nova.network.neutron [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 721.776163] env[61728]: DEBUG nova.policy [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '01f2593c9faf4c8dbccbcc9536a76a7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f22617f7e1f4598b01ebb980c61a1ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 721.805068] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b8815d13-7947-4b31-847b-24162b682e42 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "223b1f93-d33e-4566-8c8e-75234a491814" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.969s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.891122] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-463996, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.910570] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "refresh_cache-771341ed-8b8e-470c-9686-82650f5271b4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.910764] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquired lock "refresh_cache-771341ed-8b8e-470c-9686-82650f5271b4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.910926] env[61728]: DEBUG nova.network.neutron [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Forcefully refreshing network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 722.078979] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463997, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.129461] env[61728]: INFO nova.compute.manager [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Took 39.97 seconds to build instance. [ 722.138736] env[61728]: DEBUG nova.network.neutron [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Successfully created port: f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.230367] env[61728]: DEBUG nova.compute.manager [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 722.389680] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-463996, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.7766} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.389955] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] c84bfcd0-b145-4675-8b0a-5e8f94f65098/c84bfcd0-b145-4675-8b0a-5e8f94f65098.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 722.395083] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 722.395083] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01b74b5d-f425-48a5-a7b2-ca3f8f56843e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.401395] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be4050d-584a-45c0-b0e0-d23eefb8c220 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.405241] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 722.405241] env[61728]: value = "task-463998" [ 722.405241] env[61728]: _type = "Task" [ 722.405241] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.412789] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081e9a65-fe17-4ac6-b508-8ca9b14229e9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.421630] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-463998, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.454957] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129427ea-ba05-4326-93e9-684c6f737152 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.463756] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6b4177-38aa-4ae0-9634-0fe6b6350534 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.479870] env[61728]: DEBUG nova.compute.provider_tree [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 722.570972] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463997, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.630749] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fde5f921-74a1-44a2-a5a9-2f863798eee5 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "6df7b619-8cc4-4dd9-8596-22dc83234a8e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.835s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.916645] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-463998, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076232} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.916988] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 722.917814] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c888f0-ee25-47bc-acab-96c356ef614d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.942431] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] c84bfcd0-b145-4675-8b0a-5e8f94f65098/c84bfcd0-b145-4675-8b0a-5e8f94f65098.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 722.942974] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2080f83-bf26-4246-a995-71e18aef949b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.965041] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 722.965041] env[61728]: value = "task-463999" [ 722.965041] env[61728]: _type = "Task" [ 722.965041] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.977517] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-463999, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.006143] env[61728]: ERROR nova.scheduler.client.report [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [req-4624a0c7-adec-41e8-b85e-740ec4955c14] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4624a0c7-adec-41e8-b85e-740ec4955c14"}]} [ 723.025551] env[61728]: DEBUG nova.scheduler.client.report [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 723.042515] env[61728]: DEBUG nova.scheduler.client.report [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 723.042515] env[61728]: DEBUG nova.compute.provider_tree [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 723.059250] env[61728]: DEBUG nova.scheduler.client.report [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 723.074734] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463997, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.081524] env[61728]: DEBUG nova.scheduler.client.report [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 723.134193] env[61728]: DEBUG nova.compute.manager [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.171421] env[61728]: DEBUG nova.network.neutron [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Updating instance_info_cache with network_info: [{"id": "cc02ac93-777e-4bcf-8af5-69f162374735", "address": "fa:16:3e:fc:57:f3", "network": {"id": "66be9cbe-252b-4b34-9fee-3ab168f5693f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1778550721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cad2b09b986d4096ad91489a99b8d27e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc02ac93-77", "ovs_interfaceid": "cc02ac93-777e-4bcf-8af5-69f162374735", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.241871] env[61728]: DEBUG nova.compute.manager [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 723.276364] env[61728]: DEBUG nova.virt.hardware [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.276628] env[61728]: DEBUG nova.virt.hardware [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.276796] env[61728]: DEBUG nova.virt.hardware [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.276984] env[61728]: DEBUG nova.virt.hardware [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.277153] env[61728]: DEBUG nova.virt.hardware [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.277307] env[61728]: DEBUG nova.virt.hardware [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.277520] env[61728]: DEBUG nova.virt.hardware [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.277684] env[61728]: DEBUG nova.virt.hardware [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.277860] env[61728]: DEBUG nova.virt.hardware [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.278044] env[61728]: DEBUG nova.virt.hardware [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.278233] env[61728]: DEBUG nova.virt.hardware [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.279101] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79432a83-d4d5-4e97-86ab-35ae7f72ee32 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.288284] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081b468e-990e-44d4-9c49-c478a9da74b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.479159] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-463999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.545235] env[61728]: DEBUG nova.compute.manager [req-0e70ba9f-505d-4790-ae14-57324398f7e7 req-d9081f42-7fca-4bb4-a112-81ad4bf2a663 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Received event network-changed-401726ae-678d-4918-b8c1-441b3802ddc8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 723.545235] env[61728]: DEBUG nova.compute.manager [req-0e70ba9f-505d-4790-ae14-57324398f7e7 req-d9081f42-7fca-4bb4-a112-81ad4bf2a663 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Refreshing instance network info cache due to event network-changed-401726ae-678d-4918-b8c1-441b3802ddc8. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 723.545235] env[61728]: DEBUG oslo_concurrency.lockutils [req-0e70ba9f-505d-4790-ae14-57324398f7e7 req-d9081f42-7fca-4bb4-a112-81ad4bf2a663 service nova] Acquiring lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.545235] env[61728]: DEBUG oslo_concurrency.lockutils [req-0e70ba9f-505d-4790-ae14-57324398f7e7 req-d9081f42-7fca-4bb4-a112-81ad4bf2a663 service nova] Acquired lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.545235] env[61728]: DEBUG nova.network.neutron [req-0e70ba9f-505d-4790-ae14-57324398f7e7 req-d9081f42-7fca-4bb4-a112-81ad4bf2a663 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Refreshing network info cache for port 401726ae-678d-4918-b8c1-441b3802ddc8 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 723.581996] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463997, 'name': CloneVM_Task} progress is 95%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.667316] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.670183] env[61728]: DEBUG nova.network.neutron [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Successfully updated port: f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 723.677170] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Releasing lock "refresh_cache-771341ed-8b8e-470c-9686-82650f5271b4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.677170] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Updated the network info_cache for instance {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 723.677170] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.677170] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.677170] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.677351] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.677418] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.677558] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.677677] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61728) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 723.677819] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.698488] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2858d650-b7a5-4f86-8898-34f2f96603a3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.707021] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828d2e01-024e-442d-b423-db8f1d13a12a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.747804] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35abfec9-9420-4a8d-8ffa-99860a7601b2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.756918] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05954eb-621c-4855-ae13-6f537309227b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.773851] env[61728]: DEBUG nova.compute.provider_tree [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 723.815371] env[61728]: DEBUG nova.compute.manager [req-65a05487-75ba-41cf-9a64-5e54392425d4 req-eadceb79-37b4-4798-b8f4-070a76dbddb0 service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Received event network-vif-plugged-f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 723.815652] env[61728]: DEBUG oslo_concurrency.lockutils [req-65a05487-75ba-41cf-9a64-5e54392425d4 req-eadceb79-37b4-4798-b8f4-070a76dbddb0 service nova] Acquiring lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.815818] env[61728]: DEBUG oslo_concurrency.lockutils [req-65a05487-75ba-41cf-9a64-5e54392425d4 req-eadceb79-37b4-4798-b8f4-070a76dbddb0 service nova] Lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.816065] env[61728]: DEBUG oslo_concurrency.lockutils [req-65a05487-75ba-41cf-9a64-5e54392425d4 req-eadceb79-37b4-4798-b8f4-070a76dbddb0 service nova] Lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.816326] env[61728]: DEBUG nova.compute.manager [req-65a05487-75ba-41cf-9a64-5e54392425d4 req-eadceb79-37b4-4798-b8f4-070a76dbddb0 service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] No waiting events found dispatching network-vif-plugged-f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 723.816413] env[61728]: WARNING nova.compute.manager [req-65a05487-75ba-41cf-9a64-5e54392425d4 req-eadceb79-37b4-4798-b8f4-070a76dbddb0 service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Received unexpected event network-vif-plugged-f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1 for instance with vm_state building and task_state spawning. [ 723.976909] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-463999, 'name': ReconfigVM_Task, 'duration_secs': 0.93571} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.977172] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Reconfigured VM instance instance-00000033 to attach disk [datastore1] c84bfcd0-b145-4675-8b0a-5e8f94f65098/c84bfcd0-b145-4675-8b0a-5e8f94f65098.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 723.978402] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5cb070fc-d761-4a7f-a556-b0a48bd0c27b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.987941] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 723.987941] env[61728]: value = "task-464000" [ 723.987941] env[61728]: _type = "Task" [ 723.987941] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.997610] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464000, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.072779] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-463997, 'name': CloneVM_Task, 'duration_secs': 2.246919} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.073097] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Created linked-clone VM from snapshot [ 724.073821] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e61acbf7-4739-4a27-91ef-bfaf0a3363bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.082099] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Uploading image 998ee441-c19d-4804-9fda-fe842a088ddb {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 724.104661] env[61728]: DEBUG oslo_vmware.rw_handles [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 724.104661] env[61728]: value = "vm-122068" [ 724.104661] env[61728]: _type = "VirtualMachine" [ 724.104661] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 724.104979] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-46427fa7-e7de-433f-8900-1c3dc922a32e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.114299] env[61728]: DEBUG oslo_vmware.rw_handles [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lease: (returnval){ [ 724.114299] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522d1484-dc0a-f581-451e-5d37f5656cba" [ 724.114299] env[61728]: _type = "HttpNfcLease" [ 724.114299] env[61728]: } obtained for exporting VM: (result){ [ 724.114299] env[61728]: value = "vm-122068" [ 724.114299] env[61728]: _type = "VirtualMachine" [ 724.114299] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 724.114678] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the lease: (returnval){ [ 724.114678] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522d1484-dc0a-f581-451e-5d37f5656cba" [ 724.114678] env[61728]: _type = "HttpNfcLease" [ 724.114678] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 724.125035] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 724.125035] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522d1484-dc0a-f581-451e-5d37f5656cba" [ 724.125035] env[61728]: _type = "HttpNfcLease" [ 724.125035] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 724.180143] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.180143] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquired lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.180143] env[61728]: DEBUG nova.network.neutron [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 724.180795] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.289959] env[61728]: DEBUG nova.network.neutron [req-0e70ba9f-505d-4790-ae14-57324398f7e7 req-d9081f42-7fca-4bb4-a112-81ad4bf2a663 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Updated VIF entry in instance network info cache for port 401726ae-678d-4918-b8c1-441b3802ddc8. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 724.290350] env[61728]: DEBUG nova.network.neutron [req-0e70ba9f-505d-4790-ae14-57324398f7e7 req-d9081f42-7fca-4bb4-a112-81ad4bf2a663 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Updating instance_info_cache with network_info: [{"id": "401726ae-678d-4918-b8c1-441b3802ddc8", "address": "fa:16:3e:36:16:cf", "network": {"id": "5d1e643d-f8a4-4302-9920-bb661df4632d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1891316321-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fc16f1d1b144286b180cd7c8f7588fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "877ed63d-906e-4bd5-a1fc-7e82d172d41e", "external-id": "nsx-vlan-transportzone-642", "segmentation_id": 642, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap401726ae-67", "ovs_interfaceid": "401726ae-678d-4918-b8c1-441b3802ddc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.306335] env[61728]: DEBUG nova.scheduler.client.report [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 87 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 724.306335] env[61728]: DEBUG nova.compute.provider_tree [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 87 to 88 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 724.306828] env[61728]: DEBUG nova.compute.provider_tree [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 724.349018] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "223b1f93-d33e-4566-8c8e-75234a491814" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.349289] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "223b1f93-d33e-4566-8c8e-75234a491814" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.349513] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "223b1f93-d33e-4566-8c8e-75234a491814-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.350385] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "223b1f93-d33e-4566-8c8e-75234a491814-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.350587] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "223b1f93-d33e-4566-8c8e-75234a491814-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.352770] env[61728]: INFO nova.compute.manager [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Terminating instance [ 724.354685] env[61728]: DEBUG nova.compute.manager [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 724.354910] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 724.355983] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01797537-29d1-407f-ae9b-89efd39ef2ef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.364585] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 724.365140] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e3948fe4-d556-43ec-b6bb-6783b9fce28d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.373162] env[61728]: DEBUG oslo_vmware.api [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 724.373162] env[61728]: value = "task-464002" [ 724.373162] env[61728]: _type = "Task" [ 724.373162] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.382676] env[61728]: DEBUG oslo_vmware.api [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-464002, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.499710] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464000, 'name': Rename_Task, 'duration_secs': 0.16022} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.500837] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 724.500837] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5be849b4-da31-4fa1-8172-87a939a7ba4a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.510703] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 724.510703] env[61728]: value = "task-464003" [ 724.510703] env[61728]: _type = "Task" [ 724.510703] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.521356] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464003, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.623528] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 724.623528] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522d1484-dc0a-f581-451e-5d37f5656cba" [ 724.623528] env[61728]: _type = "HttpNfcLease" [ 724.623528] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 724.623528] env[61728]: DEBUG oslo_vmware.rw_handles [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 724.623528] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522d1484-dc0a-f581-451e-5d37f5656cba" [ 724.623528] env[61728]: _type = "HttpNfcLease" [ 724.623528] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 724.624230] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7f0b12-6413-4af8-9740-8149522df668 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.632728] env[61728]: DEBUG oslo_vmware.rw_handles [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524ea2d9-ed8a-9b9d-3662-60668b9be423/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 724.632958] env[61728]: DEBUG oslo_vmware.rw_handles [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524ea2d9-ed8a-9b9d-3662-60668b9be423/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 724.721867] env[61728]: DEBUG nova.network.neutron [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.723797] env[61728]: DEBUG nova.compute.manager [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 724.724755] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd293ad-31cc-44d8-b25e-8f4f5a3754ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.765126] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b7a74fb1-780a-4f54-803c-5c0342649bb1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.792598] env[61728]: DEBUG oslo_concurrency.lockutils [req-0e70ba9f-505d-4790-ae14-57324398f7e7 req-d9081f42-7fca-4bb4-a112-81ad4bf2a663 service nova] Releasing lock "refresh_cache-223b1f93-d33e-4566-8c8e-75234a491814" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.811878] env[61728]: DEBUG oslo_concurrency.lockutils [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.592s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.816583] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.576s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.816583] env[61728]: DEBUG nova.objects.instance [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Lazy-loading 'resources' on Instance uuid ad304c11-4ff8-4910-92a5-24216d727c4b {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 724.842201] env[61728]: INFO nova.scheduler.client.report [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Deleted allocations for instance 650d2875-68f7-4131-b9b7-9799ee08f1d6 [ 724.884569] env[61728]: DEBUG oslo_vmware.api [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-464002, 'name': PowerOffVM_Task, 'duration_secs': 0.196389} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.884894] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 724.885122] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 724.885404] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5ffdc1b5-784a-4551-9ff0-3d9bfdedb42a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.896066] env[61728]: DEBUG nova.network.neutron [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Updating instance_info_cache with network_info: [{"id": "f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1", "address": "fa:16:3e:f7:31:b7", "network": {"id": "105779cb-8a8c-49ab-8ea4-00d3c7878a6b", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1826811954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f22617f7e1f4598b01ebb980c61a1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9bdd9c4-3e", "ovs_interfaceid": "f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.968141] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 724.968435] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 724.968706] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Deleting the datastore file [datastore1] 223b1f93-d33e-4566-8c8e-75234a491814 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 724.969871] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a279bacb-8115-4f74-9bf3-ee018bda60d1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.978717] env[61728]: DEBUG oslo_vmware.api [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 724.978717] env[61728]: value = "task-464005" [ 724.978717] env[61728]: _type = "Task" [ 724.978717] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.990290] env[61728]: DEBUG oslo_vmware.api [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-464005, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.024946] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464003, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.237056] env[61728]: INFO nova.compute.manager [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] instance snapshotting [ 725.239561] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5dbaf98-0b69-41e8-a278-f75fa86e8bce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.261833] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f930239-7337-4f14-b76e-ad09e254ced7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.353660] env[61728]: DEBUG oslo_concurrency.lockutils [None req-80cb752c-a450-4220-acfa-1979d36b3869 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "650d2875-68f7-4131-b9b7-9799ee08f1d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.574s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.399628] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Releasing lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.400013] env[61728]: DEBUG nova.compute.manager [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Instance network_info: |[{"id": "f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1", "address": "fa:16:3e:f7:31:b7", "network": {"id": "105779cb-8a8c-49ab-8ea4-00d3c7878a6b", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1826811954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f22617f7e1f4598b01ebb980c61a1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9bdd9c4-3e", "ovs_interfaceid": "f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 725.400608] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:31:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '77aa121f-8fb6-42f3-aaea-43addfe449b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 725.409130] env[61728]: DEBUG oslo.service.loopingcall [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.412626] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 725.413339] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2e2270d3-09ed-4a24-bd1a-db2386d4a507 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.439487] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 725.439487] env[61728]: value = "task-464006" [ 725.439487] env[61728]: _type = "Task" [ 725.439487] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.454500] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464006, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.489885] env[61728]: DEBUG oslo_vmware.api [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-464005, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174465} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.492320] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 725.492609] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 725.492792] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 725.492965] env[61728]: INFO nova.compute.manager [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Took 1.14 seconds to destroy the instance on the hypervisor. [ 725.493270] env[61728]: DEBUG oslo.service.loopingcall [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.493695] env[61728]: DEBUG nova.compute.manager [-] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.493820] env[61728]: DEBUG nova.network.neutron [-] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 725.526196] env[61728]: DEBUG oslo_vmware.api [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464003, 'name': PowerOnVM_Task, 'duration_secs': 0.551481} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.526463] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 725.526696] env[61728]: INFO nova.compute.manager [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Took 8.76 seconds to spawn the instance on the hypervisor. [ 725.526941] env[61728]: DEBUG nova.compute.manager [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 725.528022] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668d22f0-d18a-4088-a7e8-8a610807ac6e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.659998] env[61728]: DEBUG oslo_concurrency.lockutils [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "c8f1f4ae-ab08-4a03-a3fd-94014509e7e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.660295] env[61728]: DEBUG oslo_concurrency.lockutils [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "c8f1f4ae-ab08-4a03-a3fd-94014509e7e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.661061] env[61728]: DEBUG oslo_concurrency.lockutils [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "c8f1f4ae-ab08-4a03-a3fd-94014509e7e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.661388] env[61728]: DEBUG oslo_concurrency.lockutils [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "c8f1f4ae-ab08-4a03-a3fd-94014509e7e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.661700] env[61728]: DEBUG oslo_concurrency.lockutils [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "c8f1f4ae-ab08-4a03-a3fd-94014509e7e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.664868] env[61728]: INFO nova.compute.manager [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Terminating instance [ 725.665851] env[61728]: DEBUG nova.compute.manager [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 725.666067] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 725.666915] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7f558f-edf2-4437-b15c-898299105c5e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.680645] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 725.681732] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70864d5e-7743-4dc3-bafd-290b7b851053 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.690174] env[61728]: DEBUG oslo_vmware.api [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 725.690174] env[61728]: value = "task-464007" [ 725.690174] env[61728]: _type = "Task" [ 725.690174] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.702973] env[61728]: DEBUG oslo_vmware.api [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-464007, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.774508] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 725.774939] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b1c3817a-e3ec-4fc7-b658-7395e7826232 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.783588] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 725.783588] env[61728]: value = "task-464008" [ 725.783588] env[61728]: _type = "Task" [ 725.783588] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.803696] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464008, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.953884] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464006, 'name': CreateVM_Task, 'duration_secs': 0.453625} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.954243] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 725.955256] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.955443] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.955785] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 725.956084] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b16e6a4-73d0-48a3-ad7b-e8e947cc674f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.967021] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 725.967021] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529627d1-3979-ac2c-020a-3443275e61d2" [ 725.967021] env[61728]: _type = "Task" [ 725.967021] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.976321] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529627d1-3979-ac2c-020a-3443275e61d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.998524] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda13b95-1d5d-4171-bb5e-0c06654f4f1a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.004954] env[61728]: DEBUG nova.compute.manager [req-9938ebb2-a9a7-48fb-8bf5-ddd0b90a7f09 req-9fd451aa-36b7-401b-8138-29d1bccc0b73 service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Received event network-changed-f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 726.005185] env[61728]: DEBUG nova.compute.manager [req-9938ebb2-a9a7-48fb-8bf5-ddd0b90a7f09 req-9fd451aa-36b7-401b-8138-29d1bccc0b73 service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Refreshing instance network info cache due to event network-changed-f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 726.005422] env[61728]: DEBUG oslo_concurrency.lockutils [req-9938ebb2-a9a7-48fb-8bf5-ddd0b90a7f09 req-9fd451aa-36b7-401b-8138-29d1bccc0b73 service nova] Acquiring lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.005606] env[61728]: DEBUG oslo_concurrency.lockutils [req-9938ebb2-a9a7-48fb-8bf5-ddd0b90a7f09 req-9fd451aa-36b7-401b-8138-29d1bccc0b73 service nova] Acquired lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.005818] env[61728]: DEBUG nova.network.neutron [req-9938ebb2-a9a7-48fb-8bf5-ddd0b90a7f09 req-9fd451aa-36b7-401b-8138-29d1bccc0b73 service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Refreshing network info cache for port f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 726.012304] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9179bc5-2439-4cbc-8818-1109476cb899 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.067678] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b89a415-247a-4ce5-85e0-15920b3cc5e4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.071692] env[61728]: INFO nova.compute.manager [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Took 42.33 seconds to build instance. [ 726.083051] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8387389-a3de-4e72-a418-64f98089eed9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.099907] env[61728]: DEBUG nova.compute.provider_tree [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.200436] env[61728]: DEBUG oslo_vmware.api [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-464007, 'name': PowerOffVM_Task, 'duration_secs': 0.223479} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.200775] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 726.201051] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 726.201352] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-68f43ede-4544-4654-96bf-b4799aaa0a61 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.272664] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 726.272664] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 726.272664] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Deleting the datastore file [datastore1] c8f1f4ae-ab08-4a03-a3fd-94014509e7e7 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 726.272664] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9c25d2c6-bd82-4cf7-95c8-9a47de1ddb6d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.281400] env[61728]: DEBUG oslo_vmware.api [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for the task: (returnval){ [ 726.281400] env[61728]: value = "task-464010" [ 726.281400] env[61728]: _type = "Task" [ 726.281400] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.291862] env[61728]: DEBUG oslo_vmware.api [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-464010, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.300807] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464008, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.329909] env[61728]: DEBUG nova.network.neutron [-] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.477915] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529627d1-3979-ac2c-020a-3443275e61d2, 'name': SearchDatastore_Task, 'duration_secs': 0.018142} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.480115] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.480115] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 726.480115] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.480115] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.480115] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 726.480115] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cf662efa-01cb-4b75-a7dc-fbd008deeb06 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.493271] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 726.493538] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 726.495268] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Acquiring lock "8c4b7227-1a01-4400-beb8-2df27e17e329" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.495564] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Lock "8c4b7227-1a01-4400-beb8-2df27e17e329" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.496818] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-417386e7-aa6d-4cf7-a1e7-d1f67784ddd9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.505055] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 726.505055] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f615db-712d-ee26-6e1f-5d09f2785250" [ 726.505055] env[61728]: _type = "Task" [ 726.505055] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.517256] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f615db-712d-ee26-6e1f-5d09f2785250, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.574039] env[61728]: DEBUG oslo_concurrency.lockutils [None req-190d17c3-6bdf-47f2-9cac-9ff52ca81e8a tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.946s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.607020] env[61728]: DEBUG nova.scheduler.client.report [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.775902] env[61728]: DEBUG nova.network.neutron [req-9938ebb2-a9a7-48fb-8bf5-ddd0b90a7f09 req-9fd451aa-36b7-401b-8138-29d1bccc0b73 service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Updated VIF entry in instance network info cache for port f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 726.776329] env[61728]: DEBUG nova.network.neutron [req-9938ebb2-a9a7-48fb-8bf5-ddd0b90a7f09 req-9fd451aa-36b7-401b-8138-29d1bccc0b73 service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Updating instance_info_cache with network_info: [{"id": "f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1", "address": "fa:16:3e:f7:31:b7", "network": {"id": "105779cb-8a8c-49ab-8ea4-00d3c7878a6b", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1826811954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f22617f7e1f4598b01ebb980c61a1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9bdd9c4-3e", "ovs_interfaceid": "f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.794760] env[61728]: DEBUG oslo_vmware.api [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Task: {'id': task-464010, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191862} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.799212] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 726.799556] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 726.799787] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 726.800029] env[61728]: INFO nova.compute.manager [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Took 1.13 seconds to destroy the instance on the hypervisor. [ 726.800464] env[61728]: DEBUG oslo.service.loopingcall [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.801998] env[61728]: DEBUG nova.compute.manager [-] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.801998] env[61728]: DEBUG nova.network.neutron [-] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 726.810534] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464008, 'name': CreateSnapshot_Task, 'duration_secs': 0.802718} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.810855] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 726.811785] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9159e4e5-050d-40d4-b9a5-bb5a72e56005 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.834419] env[61728]: INFO nova.compute.manager [-] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Took 1.34 seconds to deallocate network for instance. [ 727.001229] env[61728]: DEBUG nova.compute.utils [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 727.020335] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f615db-712d-ee26-6e1f-5d09f2785250, 'name': SearchDatastore_Task, 'duration_secs': 0.01528} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.024903] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32b94f09-bafe-485c-8e73-af07ba112cc3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.032227] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 727.032227] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5201564e-b45d-ab77-99ae-f856c9c3cfa6" [ 727.032227] env[61728]: _type = "Task" [ 727.032227] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.044019] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5201564e-b45d-ab77-99ae-f856c9c3cfa6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.067348] env[61728]: DEBUG nova.compute.manager [req-55b0df03-3398-4ef2-aeed-3a0b4ba1d27c req-a12022be-0253-4d4d-8235-7905ddafcf4d service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Received event network-changed-372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 727.067513] env[61728]: DEBUG nova.compute.manager [req-55b0df03-3398-4ef2-aeed-3a0b4ba1d27c req-a12022be-0253-4d4d-8235-7905ddafcf4d service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Refreshing instance network info cache due to event network-changed-372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 727.067738] env[61728]: DEBUG oslo_concurrency.lockutils [req-55b0df03-3398-4ef2-aeed-3a0b4ba1d27c req-a12022be-0253-4d4d-8235-7905ddafcf4d service nova] Acquiring lock "refresh_cache-c84bfcd0-b145-4675-8b0a-5e8f94f65098" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.067864] env[61728]: DEBUG oslo_concurrency.lockutils [req-55b0df03-3398-4ef2-aeed-3a0b4ba1d27c req-a12022be-0253-4d4d-8235-7905ddafcf4d service nova] Acquired lock "refresh_cache-c84bfcd0-b145-4675-8b0a-5e8f94f65098" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.068185] env[61728]: DEBUG nova.network.neutron [req-55b0df03-3398-4ef2-aeed-3a0b4ba1d27c req-a12022be-0253-4d4d-8235-7905ddafcf4d service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Refreshing network info cache for port 372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 727.084253] env[61728]: DEBUG nova.compute.manager [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 727.111354] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.297s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.115019] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.184s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.116251] env[61728]: INFO nova.compute.claims [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.138025] env[61728]: INFO nova.scheduler.client.report [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Deleted allocations for instance ad304c11-4ff8-4910-92a5-24216d727c4b [ 727.279718] env[61728]: DEBUG oslo_concurrency.lockutils [req-9938ebb2-a9a7-48fb-8bf5-ddd0b90a7f09 req-9fd451aa-36b7-401b-8138-29d1bccc0b73 service nova] Releasing lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.280080] env[61728]: DEBUG nova.compute.manager [req-9938ebb2-a9a7-48fb-8bf5-ddd0b90a7f09 req-9fd451aa-36b7-401b-8138-29d1bccc0b73 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Received event network-vif-deleted-401726ae-678d-4918-b8c1-441b3802ddc8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 727.280286] env[61728]: INFO nova.compute.manager [req-9938ebb2-a9a7-48fb-8bf5-ddd0b90a7f09 req-9fd451aa-36b7-401b-8138-29d1bccc0b73 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Neutron deleted interface 401726ae-678d-4918-b8c1-441b3802ddc8; detaching it from the instance and deleting it from the info cache [ 727.280475] env[61728]: DEBUG nova.network.neutron [req-9938ebb2-a9a7-48fb-8bf5-ddd0b90a7f09 req-9fd451aa-36b7-401b-8138-29d1bccc0b73 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.332379] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 727.333125] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f484a5ef-e562-48e4-a470-69db25edbaf3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.343962] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.344352] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 727.344352] env[61728]: value = "task-464011" [ 727.344352] env[61728]: _type = "Task" [ 727.344352] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.353245] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464011, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.506303] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Lock "8c4b7227-1a01-4400-beb8-2df27e17e329" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.544536] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5201564e-b45d-ab77-99ae-f856c9c3cfa6, 'name': SearchDatastore_Task, 'duration_secs': 0.015779} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.544966] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.545255] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48/3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 727.545533] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d530fb64-c1e7-4c1c-aac2-85aae75827ef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.558465] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 727.558465] env[61728]: value = "task-464012" [ 727.558465] env[61728]: _type = "Task" [ 727.558465] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.570337] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464012, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.574554] env[61728]: DEBUG nova.network.neutron [-] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.604436] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.646191] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ab03e6fe-a7df-4911-a4ef-be9e49779969 tempest-ServerMetadataNegativeTestJSON-1593019073 tempest-ServerMetadataNegativeTestJSON-1593019073-project-member] Lock "ad304c11-4ff8-4910-92a5-24216d727c4b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.329s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.786610] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-825dadaa-d060-4c17-9744-5eec78d37899 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.802827] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d3d4d3-1911-4460-984b-357cac0781c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.837270] env[61728]: DEBUG nova.network.neutron [req-55b0df03-3398-4ef2-aeed-3a0b4ba1d27c req-a12022be-0253-4d4d-8235-7905ddafcf4d service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Updated VIF entry in instance network info cache for port 372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 727.837801] env[61728]: DEBUG nova.network.neutron [req-55b0df03-3398-4ef2-aeed-3a0b4ba1d27c req-a12022be-0253-4d4d-8235-7905ddafcf4d service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Updating instance_info_cache with network_info: [{"id": "372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8", "address": "fa:16:3e:68:e7:ba", "network": {"id": "22e73f4c-99d9-4e65-8a72-f6ea53d88eaf", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2136421023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "29dfac4a253f43be811df957e0ee5f96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap372c7b2a-4d", "ovs_interfaceid": "372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.857161] env[61728]: DEBUG nova.compute.manager [req-9938ebb2-a9a7-48fb-8bf5-ddd0b90a7f09 req-9fd451aa-36b7-401b-8138-29d1bccc0b73 service nova] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Detach interface failed, port_id=401726ae-678d-4918-b8c1-441b3802ddc8, reason: Instance 223b1f93-d33e-4566-8c8e-75234a491814 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 727.858879] env[61728]: DEBUG oslo_concurrency.lockutils [req-55b0df03-3398-4ef2-aeed-3a0b4ba1d27c req-a12022be-0253-4d4d-8235-7905ddafcf4d service nova] Releasing lock "refresh_cache-c84bfcd0-b145-4675-8b0a-5e8f94f65098" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.870823] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464011, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.074547] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464012, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.078397] env[61728]: INFO nova.compute.manager [-] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Took 1.28 seconds to deallocate network for instance. [ 728.281444] env[61728]: DEBUG nova.compute.manager [req-970b62e6-2310-4cfd-80a4-e30f35540a99 req-4fad65a5-3e43-41a1-b3d6-4b3e3de979df service nova] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Received event network-vif-deleted-36a914b9-8961-48e0-b44c-25f49ddbade5 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 728.370282] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464011, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.571612] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464012, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.558766} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.571982] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48/3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 728.572218] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 728.572485] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-74e76f65-0b16-4e58-9689-4506a6354bbb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.585563] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 728.585563] env[61728]: value = "task-464013" [ 728.585563] env[61728]: _type = "Task" [ 728.585563] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.586566] env[61728]: DEBUG oslo_concurrency.lockutils [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.593924] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Acquiring lock "8c4b7227-1a01-4400-beb8-2df27e17e329" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.594241] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Lock "8c4b7227-1a01-4400-beb8-2df27e17e329" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.594475] env[61728]: INFO nova.compute.manager [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Attaching volume bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7 to /dev/sdb [ 728.609335] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464013, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.645233] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3785794e-6c32-4422-9082-352a33b6313c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.653461] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be3a6809-4c44-47db-9b20-0546cec085af {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.668861] env[61728]: DEBUG nova.virt.block_device [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Updating existing volume attachment record: 3cc50b7f-3069-4904-b39c-7a0d1e37f073 {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 728.712040] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d123a6a4-a63c-4db8-bc33-aacf4cd5863a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.719651] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05cb60c7-8091-4dd5-a463-14802291864b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.758961] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b13935-3576-41f4-a33b-fc1ffb65e392 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.767766] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2898bc2d-84a9-430b-b7c9-92aa2d742ab5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.782868] env[61728]: DEBUG nova.compute.provider_tree [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.867997] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464011, 'name': CloneVM_Task} progress is 95%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.097946] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464013, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075468} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.098856] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 729.099662] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7155f88a-ce47-4103-95a4-071e87774a0f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.132977] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48/3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 729.132977] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40ff3d47-e78b-422c-a6d0-10fd1acb69c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.164707] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 729.164707] env[61728]: value = "task-464017" [ 729.164707] env[61728]: _type = "Task" [ 729.164707] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.174846] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464017, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.286049] env[61728]: DEBUG nova.scheduler.client.report [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.369160] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464011, 'name': CloneVM_Task, 'duration_secs': 1.980515} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.369439] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Created linked-clone VM from snapshot [ 729.370249] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a366d2-1105-42c3-a93e-e897dc9d09e2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.380958] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Uploading image 7ac64905-f9d8-4de1-beb4-cd6197d96f1e {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 729.411407] env[61728]: DEBUG oslo_vmware.rw_handles [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 729.411407] env[61728]: value = "vm-122071" [ 729.411407] env[61728]: _type = "VirtualMachine" [ 729.411407] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 729.411407] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3f194d1f-b946-4d44-bf8d-1c562e531060 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.422344] env[61728]: DEBUG oslo_vmware.rw_handles [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lease: (returnval){ [ 729.422344] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e6031f-5576-48d3-0f73-2dd8ae4da296" [ 729.422344] env[61728]: _type = "HttpNfcLease" [ 729.422344] env[61728]: } obtained for exporting VM: (result){ [ 729.422344] env[61728]: value = "vm-122071" [ 729.422344] env[61728]: _type = "VirtualMachine" [ 729.422344] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 729.422344] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the lease: (returnval){ [ 729.422344] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e6031f-5576-48d3-0f73-2dd8ae4da296" [ 729.422344] env[61728]: _type = "HttpNfcLease" [ 729.422344] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 729.430329] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 729.430329] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e6031f-5576-48d3-0f73-2dd8ae4da296" [ 729.430329] env[61728]: _type = "HttpNfcLease" [ 729.430329] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 729.677867] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464017, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.793202] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.679s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.793929] env[61728]: DEBUG nova.compute.manager [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 729.797044] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.118s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.797224] env[61728]: DEBUG nova.objects.instance [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lazy-loading 'resources' on Instance uuid 507b5aba-aec9-438b-92c0-5efa17959bc7 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 729.933131] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 729.933131] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e6031f-5576-48d3-0f73-2dd8ae4da296" [ 729.933131] env[61728]: _type = "HttpNfcLease" [ 729.933131] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 729.933131] env[61728]: DEBUG oslo_vmware.rw_handles [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 729.933131] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e6031f-5576-48d3-0f73-2dd8ae4da296" [ 729.933131] env[61728]: _type = "HttpNfcLease" [ 729.933131] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 729.933131] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9454e404-d116-43ac-81b5-da3bcdd0b5cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.942939] env[61728]: DEBUG oslo_vmware.rw_handles [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524bf492-0faa-abc5-7153-3352deadaf5b/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 729.944022] env[61728]: DEBUG oslo_vmware.rw_handles [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524bf492-0faa-abc5-7153-3352deadaf5b/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 730.087297] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8b9e0304-10ea-4ef6-91ef-b8b1ff43bb28 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.179031] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464017, 'name': ReconfigVM_Task, 'duration_secs': 0.625529} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.179031] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48/3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 730.179418] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ca4e94d-e82d-49c5-83ce-c9fd57e29f1e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.188290] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 730.188290] env[61728]: value = "task-464019" [ 730.188290] env[61728]: _type = "Task" [ 730.188290] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.202718] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464019, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.303959] env[61728]: DEBUG nova.compute.utils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 730.305718] env[61728]: DEBUG nova.compute.manager [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 730.305903] env[61728]: DEBUG nova.network.neutron [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 730.425311] env[61728]: DEBUG nova.policy [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '38f64fd31712472a841f82c620d9340b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f5329f59d8bf43b2a855c4259387fccd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 730.707848] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464019, 'name': Rename_Task, 'duration_secs': 0.422471} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.709644] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 730.709644] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-758dbd27-1edb-486c-9611-1d840a745c05 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.717816] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 730.717816] env[61728]: value = "task-464020" [ 730.717816] env[61728]: _type = "Task" [ 730.717816] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.735037] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464020, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.812837] env[61728]: DEBUG nova.compute.manager [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 730.895485] env[61728]: DEBUG nova.network.neutron [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Successfully created port: cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 731.138020] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21bc093a-7782-44e8-a335-8afceb741372 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.146031] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d54bb3b-1625-4255-b44b-8661385760b1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.182978] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6c9b7d-da59-4d56-a42e-9fe1d1c3dec3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.193271] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-483cb08d-124e-4280-982f-44b4bc66f5f0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.213750] env[61728]: DEBUG nova.compute.provider_tree [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.236553] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464020, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.717519] env[61728]: DEBUG nova.scheduler.client.report [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.736763] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464020, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.824954] env[61728]: DEBUG nova.compute.manager [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 732.223602] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.427s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.227089] env[61728]: DEBUG oslo_concurrency.lockutils [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 36.682s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.228756] env[61728]: INFO nova.compute.claims [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.244068] env[61728]: DEBUG oslo_vmware.api [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464020, 'name': PowerOnVM_Task, 'duration_secs': 1.484531} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.244419] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 732.244624] env[61728]: INFO nova.compute.manager [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Took 9.00 seconds to spawn the instance on the hypervisor. [ 732.244811] env[61728]: DEBUG nova.compute.manager [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 732.245655] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba21b980-e037-4ce9-87f0-148f5b7e56cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.258344] env[61728]: INFO nova.scheduler.client.report [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleted allocations for instance 507b5aba-aec9-438b-92c0-5efa17959bc7 [ 732.560930] env[61728]: DEBUG nova.network.neutron [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Successfully updated port: cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 732.771580] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8542820e-f79b-486a-9ed6-0589a137390f tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "507b5aba-aec9-438b-92c0-5efa17959bc7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 44.813s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.772632] env[61728]: INFO nova.compute.manager [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Took 46.37 seconds to build instance. [ 733.067413] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.067631] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.067793] env[61728]: DEBUG nova.network.neutron [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 733.250762] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 733.251109] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122074', 'volume_id': 'bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7', 'name': 'volume-bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8c4b7227-1a01-4400-beb8-2df27e17e329', 'attached_at': '', 'detached_at': '', 'volume_id': 'bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7', 'serial': 'bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 733.252156] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6b0345-4a7f-4b87-b61f-b75b997753e0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.272569] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4f40fc-0642-4cb5-b73c-9f2132d6d7e7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.278093] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1e29270-6c6c-47aa-b161-33ee01cfdb44 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.575s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.303394] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Reconfiguring VM instance instance-0000000f to attach disk [datastore1] volume-bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7/volume-bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 733.308133] env[61728]: DEBUG nova.compute.manager [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 733.309857] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e0b2b5f-e285-4162-98a0-066fa277feda {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.335417] env[61728]: DEBUG oslo_vmware.api [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Waiting for the task: (returnval){ [ 733.335417] env[61728]: value = "task-464022" [ 733.335417] env[61728]: _type = "Task" [ 733.335417] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.345614] env[61728]: DEBUG oslo_vmware.api [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Task: {'id': task-464022, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.605689] env[61728]: DEBUG nova.network.neutron [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.764145] env[61728]: DEBUG nova.network.neutron [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating instance_info_cache with network_info: [{"id": "cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd", "address": "fa:16:3e:fc:46:55", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc2e7ae1-f4", "ovs_interfaceid": "cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.847227] env[61728]: DEBUG oslo_vmware.api [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Task: {'id': task-464022, 'name': ReconfigVM_Task, 'duration_secs': 0.503568} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.849359] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.849745] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Reconfigured VM instance instance-0000000f to attach disk [datastore1] volume-bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7/volume-bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 733.856719] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60dedda-f7cc-485f-80ac-4c4242911554 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.862282] env[61728]: DEBUG nova.virt.hardware [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 733.863748] env[61728]: DEBUG nova.virt.hardware [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 733.863748] env[61728]: DEBUG nova.virt.hardware [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 733.863748] env[61728]: DEBUG nova.virt.hardware [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 733.863748] env[61728]: DEBUG nova.virt.hardware [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 733.863748] env[61728]: DEBUG nova.virt.hardware [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 733.863748] env[61728]: DEBUG nova.virt.hardware [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 733.864231] env[61728]: DEBUG nova.virt.hardware [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 733.864231] env[61728]: DEBUG nova.virt.hardware [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 733.864231] env[61728]: DEBUG nova.virt.hardware [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 733.864458] env[61728]: DEBUG nova.virt.hardware [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 733.864719] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-789b280f-d644-426d-a4da-4fe3c1728b85 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.878375] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ff0839-ea38-4bcb-a032-3bea469ccf21 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.892181] env[61728]: DEBUG oslo_vmware.rw_handles [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524ea2d9-ed8a-9b9d-3662-60668b9be423/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 733.898019] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28933b9b-bf06-4b44-a247-1fa6fc5d302a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.898019] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643342f1-5d0d-473a-b912-2b9e2fcb5517 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.900872] env[61728]: DEBUG oslo_vmware.api [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Waiting for the task: (returnval){ [ 733.900872] env[61728]: value = "task-464023" [ 733.900872] env[61728]: _type = "Task" [ 733.900872] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.903122] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6f4ad4-6ff7-4eb6-87d2-53aa9cc330e3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.936311] env[61728]: DEBUG oslo_vmware.rw_handles [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524ea2d9-ed8a-9b9d-3662-60668b9be423/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 733.936532] env[61728]: ERROR oslo_vmware.rw_handles [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524ea2d9-ed8a-9b9d-3662-60668b9be423/disk-0.vmdk due to incomplete transfer. [ 733.937647] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ffb1b54-25b5-4122-ab5c-f71ccb835f73 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.944216] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-db757324-b1fc-47f9-b2fa-3a7101ba51fa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.959719] env[61728]: DEBUG oslo_vmware.api [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Task: {'id': task-464023, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.966404] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa1a92a-09f2-4f46-aeb2-c0b038b9fa2d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.972502] env[61728]: DEBUG oslo_vmware.rw_handles [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524ea2d9-ed8a-9b9d-3662-60668b9be423/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 733.972502] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Uploaded image 998ee441-c19d-4804-9fda-fe842a088ddb to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 733.973880] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 733.974518] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ff994a6f-4c9c-4b08-b391-18366433b1cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.984719] env[61728]: DEBUG nova.compute.provider_tree [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 733.987264] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 733.987264] env[61728]: value = "task-464024" [ 733.987264] env[61728]: _type = "Task" [ 733.987264] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.997347] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464024, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.267164] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.267515] env[61728]: DEBUG nova.compute.manager [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Instance network_info: |[{"id": "cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd", "address": "fa:16:3e:fc:46:55", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc2e7ae1-f4", "ovs_interfaceid": "cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 734.267975] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:46:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c492f5cc-7ae0-4cab-823c-0d5dd8c60b26', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 734.275744] env[61728]: DEBUG oslo.service.loopingcall [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.275991] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 734.276243] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-78672743-adaf-40a0-a20e-cf88054fb8b6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.298897] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 734.298897] env[61728]: value = "task-464025" [ 734.298897] env[61728]: _type = "Task" [ 734.298897] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.308691] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464025, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.421064] env[61728]: DEBUG oslo_vmware.api [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Task: {'id': task-464023, 'name': ReconfigVM_Task, 'duration_secs': 0.15409} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.421426] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122074', 'volume_id': 'bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7', 'name': 'volume-bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8c4b7227-1a01-4400-beb8-2df27e17e329', 'attached_at': '', 'detached_at': '', 'volume_id': 'bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7', 'serial': 'bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 734.504018] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464024, 'name': Destroy_Task, 'duration_secs': 0.34977} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.504018] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Destroyed the VM [ 734.504018] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 734.504018] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c40dde58-8eaa-4611-acd3-5955cfab5965 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.512439] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 734.512439] env[61728]: value = "task-464026" [ 734.512439] env[61728]: _type = "Task" [ 734.512439] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.522282] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464026, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.535791] env[61728]: DEBUG nova.scheduler.client.report [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 88 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 734.536078] env[61728]: DEBUG nova.compute.provider_tree [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 88 to 89 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 734.536319] env[61728]: DEBUG nova.compute.provider_tree [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 734.695958] env[61728]: DEBUG nova.compute.manager [req-3768a593-e003-46db-9f08-0fc43c71b990 req-684996cc-7b25-4222-a12a-c015702f0b54 service nova] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Received event network-vif-plugged-cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 734.696176] env[61728]: DEBUG oslo_concurrency.lockutils [req-3768a593-e003-46db-9f08-0fc43c71b990 req-684996cc-7b25-4222-a12a-c015702f0b54 service nova] Acquiring lock "f2692e22-6c14-4bb3-a080-607f0731105d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.696455] env[61728]: DEBUG oslo_concurrency.lockutils [req-3768a593-e003-46db-9f08-0fc43c71b990 req-684996cc-7b25-4222-a12a-c015702f0b54 service nova] Lock "f2692e22-6c14-4bb3-a080-607f0731105d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.696684] env[61728]: DEBUG oslo_concurrency.lockutils [req-3768a593-e003-46db-9f08-0fc43c71b990 req-684996cc-7b25-4222-a12a-c015702f0b54 service nova] Lock "f2692e22-6c14-4bb3-a080-607f0731105d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.696896] env[61728]: DEBUG nova.compute.manager [req-3768a593-e003-46db-9f08-0fc43c71b990 req-684996cc-7b25-4222-a12a-c015702f0b54 service nova] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] No waiting events found dispatching network-vif-plugged-cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 734.697128] env[61728]: WARNING nova.compute.manager [req-3768a593-e003-46db-9f08-0fc43c71b990 req-684996cc-7b25-4222-a12a-c015702f0b54 service nova] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Received unexpected event network-vif-plugged-cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd for instance with vm_state building and task_state spawning. [ 734.811522] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464025, 'name': CreateVM_Task, 'duration_secs': 0.401099} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.811804] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 734.812626] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.812851] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.813280] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 734.813598] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66ee5db4-a960-4774-a7b6-8e0e094ad366 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.821673] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 734.821673] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d2362b-950e-0e3f-8a32-333771be7141" [ 734.821673] env[61728]: _type = "Task" [ 734.821673] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.834304] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d2362b-950e-0e3f-8a32-333771be7141, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.026021] env[61728]: DEBUG oslo_vmware.api [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464026, 'name': RemoveSnapshot_Task, 'duration_secs': 0.472348} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.026559] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 735.026942] env[61728]: INFO nova.compute.manager [None req-594e801d-7b75-4241-87ee-7a97169d6991 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Took 15.58 seconds to snapshot the instance on the hypervisor. [ 735.045542] env[61728]: DEBUG oslo_concurrency.lockutils [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.815s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.045542] env[61728]: DEBUG nova.compute.manager [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 735.046696] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 39.396s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.336236] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d2362b-950e-0e3f-8a32-333771be7141, 'name': SearchDatastore_Task, 'duration_secs': 0.013131} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.336601] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.336729] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 735.337090] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.337283] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.337509] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 735.340077] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-323c26ce-7e3d-4607-82f4-29b7c3ac30aa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.349447] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 735.350793] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 735.350793] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4399b353-8d37-45a4-99f9-a5cc2caafff5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.358238] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 735.358238] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5281c855-b973-033f-5dd3-cffa11038c63" [ 735.358238] env[61728]: _type = "Task" [ 735.358238] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.368249] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5281c855-b973-033f-5dd3-cffa11038c63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.372792] env[61728]: INFO nova.compute.manager [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Rebuilding instance [ 735.413792] env[61728]: DEBUG nova.compute.manager [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 735.414809] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb92a8d-2dc9-4375-ad35-e242bcbd2b16 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.477651] env[61728]: DEBUG nova.objects.instance [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Lazy-loading 'flavor' on Instance uuid 8c4b7227-1a01-4400-beb8-2df27e17e329 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 735.548287] env[61728]: DEBUG nova.compute.utils [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.549650] env[61728]: DEBUG nova.compute.manager [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Not allocating networking since 'none' was specified. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 735.552647] env[61728]: INFO nova.compute.claims [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.826558] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquiring lock "b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.826829] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Lock "b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.872047] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5281c855-b973-033f-5dd3-cffa11038c63, 'name': SearchDatastore_Task, 'duration_secs': 0.013801} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.872866] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2904271-d1b8-463b-833c-87f45ada714b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.880143] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 735.880143] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f42677-b1e3-d7d2-c9c9-205d9f829355" [ 735.880143] env[61728]: _type = "Task" [ 735.880143] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.890588] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f42677-b1e3-d7d2-c9c9-205d9f829355, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.928900] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 735.929297] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ee6dde5-15d3-49b7-970c-4584f6624d4f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.938720] env[61728]: DEBUG oslo_vmware.api [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Waiting for the task: (returnval){ [ 735.938720] env[61728]: value = "task-464027" [ 735.938720] env[61728]: _type = "Task" [ 735.938720] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.954063] env[61728]: DEBUG oslo_vmware.api [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-464027, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.986850] env[61728]: DEBUG oslo_concurrency.lockutils [None req-15a6c9e3-63bc-490b-955a-937395cb6a81 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Lock "8c4b7227-1a01-4400-beb8-2df27e17e329" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.392s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.063624] env[61728]: DEBUG nova.compute.manager [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 736.069241] env[61728]: INFO nova.compute.resource_tracker [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating resource usage from migration 9939d541-20e8-411b-981f-657861975b29 [ 736.393030] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f42677-b1e3-d7d2-c9c9-205d9f829355, 'name': SearchDatastore_Task, 'duration_secs': 0.022941} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.396162] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.396495] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] f2692e22-6c14-4bb3-a080-607f0731105d/f2692e22-6c14-4bb3-a080-607f0731105d.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 736.396953] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-225176d4-55c5-4db2-96e6-50b726ab2744 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.406056] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 736.406056] env[61728]: value = "task-464028" [ 736.406056] env[61728]: _type = "Task" [ 736.406056] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.416481] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464028, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.452526] env[61728]: DEBUG oslo_vmware.api [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-464027, 'name': PowerOffVM_Task, 'duration_secs': 0.241399} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.452925] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 736.453863] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 736.457584] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-951e858a-f714-408d-9e79-2a62d990071d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.468430] env[61728]: DEBUG oslo_vmware.api [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Waiting for the task: (returnval){ [ 736.468430] env[61728]: value = "task-464029" [ 736.468430] env[61728]: _type = "Task" [ 736.468430] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.483796] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 736.484076] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 736.484446] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-121946', 'volume_id': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'name': 'volume-2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0f75629b-bdec-4593-aceb-cd0478972ccc', 'attached_at': '', 'detached_at': '', 'volume_id': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'serial': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 736.486015] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f95a9e-a0cf-4ec8-b13b-d2a0aef0c228 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.511454] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d29a5c8-beae-4a51-b41e-98bf4ce08e2c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.521040] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a48d24f-b852-4057-9d95-8526daaa180c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.545102] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827d2775-6e48-41fc-829b-f26192226953 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.562242] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] The volume has not been displaced from its original location: [datastore1] volume-2eb2f0fc-4130-44f7-bddb-eb01dbadccc4/volume-2eb2f0fc-4130-44f7-bddb-eb01dbadccc4.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 736.567681] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Reconfiguring VM instance instance-0000001d to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 736.570816] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d6d334b-35c7-4009-b56e-f48cbf63d819 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.594812] env[61728]: DEBUG oslo_vmware.api [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Waiting for the task: (returnval){ [ 736.594812] env[61728]: value = "task-464030" [ 736.594812] env[61728]: _type = "Task" [ 736.594812] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.609508] env[61728]: DEBUG oslo_vmware.api [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-464030, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.749955] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8087b501-c0c3-48a2-b84b-1467dcadf599 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.758396] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d60df8-38be-45c3-a328-872f03f7a52f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.790945] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cfae5ad-77ed-4c9b-b142-148225a1796c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.801018] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a117e7-42c8-4c7f-8d06-3f4c4e92db19 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.817990] env[61728]: DEBUG nova.compute.provider_tree [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 736.820995] env[61728]: DEBUG nova.compute.manager [req-a0db059d-8928-457c-825d-b955e9deb2e3 req-71b04c35-0497-49a8-b598-dc17209901a9 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Received event network-changed-0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 736.821222] env[61728]: DEBUG nova.compute.manager [req-a0db059d-8928-457c-825d-b955e9deb2e3 req-71b04c35-0497-49a8-b598-dc17209901a9 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Refreshing instance network info cache due to event network-changed-0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 736.821460] env[61728]: DEBUG oslo_concurrency.lockutils [req-a0db059d-8928-457c-825d-b955e9deb2e3 req-71b04c35-0497-49a8-b598-dc17209901a9 service nova] Acquiring lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.821685] env[61728]: DEBUG oslo_concurrency.lockutils [req-a0db059d-8928-457c-825d-b955e9deb2e3 req-71b04c35-0497-49a8-b598-dc17209901a9 service nova] Acquired lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.821867] env[61728]: DEBUG nova.network.neutron [req-a0db059d-8928-457c-825d-b955e9deb2e3 req-71b04c35-0497-49a8-b598-dc17209901a9 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Refreshing network info cache for port 0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 736.922324] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464028, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.089376] env[61728]: DEBUG nova.compute.manager [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 737.108141] env[61728]: DEBUG oslo_vmware.api [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-464030, 'name': ReconfigVM_Task, 'duration_secs': 0.176061} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.108476] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Reconfigured VM instance instance-0000001d to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 737.115381] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1249a4b-f27d-44a6-a015-92e9f58dc797 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.133848] env[61728]: DEBUG nova.virt.hardware [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 737.134204] env[61728]: DEBUG nova.virt.hardware [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 737.134387] env[61728]: DEBUG nova.virt.hardware [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 737.134636] env[61728]: DEBUG nova.virt.hardware [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 737.134811] env[61728]: DEBUG nova.virt.hardware [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 737.135592] env[61728]: DEBUG nova.virt.hardware [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 737.135592] env[61728]: DEBUG nova.virt.hardware [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 737.135592] env[61728]: DEBUG nova.virt.hardware [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 737.136149] env[61728]: DEBUG nova.virt.hardware [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 737.136600] env[61728]: DEBUG nova.virt.hardware [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 737.136990] env[61728]: DEBUG nova.virt.hardware [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.138040] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19567c3-aaff-4e72-abc7-e6a054869773 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.142917] env[61728]: DEBUG oslo_vmware.api [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Waiting for the task: (returnval){ [ 737.142917] env[61728]: value = "task-464031" [ 737.142917] env[61728]: _type = "Task" [ 737.142917] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.151727] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c6f2f0-21cb-4efe-9e08-c7e1312addb6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.163543] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49c8b81b-62cf-43b4-9473-273d39065e09 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "fb962a28-c923-441f-9564-814c6bcb7cab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.163806] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49c8b81b-62cf-43b4-9473-273d39065e09 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "fb962a28-c923-441f-9564-814c6bcb7cab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.164047] env[61728]: DEBUG oslo_vmware.api [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-464031, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.177579] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Instance VIF info [] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 737.183632] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Creating folder: Project (50f7ae37fa5a42df8218f5141ddf32a6). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 737.183632] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-980d0c94-1ca9-4e08-aa1a-71b66d3c6b6a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.195762] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Created folder: Project (50f7ae37fa5a42df8218f5141ddf32a6) in parent group-v121913. [ 737.195982] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Creating folder: Instances. Parent ref: group-v122076. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 737.196249] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c8f2bd63-f273-457c-bf4d-b943e026725e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.208931] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Created folder: Instances in parent group-v122076. [ 737.208931] env[61728]: DEBUG oslo.service.loopingcall [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.208931] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 737.208931] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-57564d43-c853-4279-9748-cf6f39a07816 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.226045] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 737.226045] env[61728]: value = "task-464034" [ 737.226045] env[61728]: _type = "Task" [ 737.226045] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.236818] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464034, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.331551] env[61728]: DEBUG nova.compute.manager [req-5f0c99ef-6437-44b6-9cab-bc8ec644ecbb req-92ea3252-f91c-4acb-b98f-3fcfe43e00f2 service nova] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Received event network-changed-cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 737.331916] env[61728]: DEBUG nova.compute.manager [req-5f0c99ef-6437-44b6-9cab-bc8ec644ecbb req-92ea3252-f91c-4acb-b98f-3fcfe43e00f2 service nova] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Refreshing instance network info cache due to event network-changed-cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 737.332067] env[61728]: DEBUG oslo_concurrency.lockutils [req-5f0c99ef-6437-44b6-9cab-bc8ec644ecbb req-92ea3252-f91c-4acb-b98f-3fcfe43e00f2 service nova] Acquiring lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.332276] env[61728]: DEBUG oslo_concurrency.lockutils [req-5f0c99ef-6437-44b6-9cab-bc8ec644ecbb req-92ea3252-f91c-4acb-b98f-3fcfe43e00f2 service nova] Acquired lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.332492] env[61728]: DEBUG nova.network.neutron [req-5f0c99ef-6437-44b6-9cab-bc8ec644ecbb req-92ea3252-f91c-4acb-b98f-3fcfe43e00f2 service nova] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Refreshing network info cache for port cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 737.349661] env[61728]: ERROR nova.scheduler.client.report [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [req-0cf23d35-c31a-419b-a13b-295f86af6c31] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-0cf23d35-c31a-419b-a13b-295f86af6c31"}]} [ 737.355945] env[61728]: DEBUG oslo_concurrency.lockutils [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "c14034b5-330b-4fb1-af31-d530e3231992" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.356215] env[61728]: DEBUG oslo_concurrency.lockutils [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "c14034b5-330b-4fb1-af31-d530e3231992" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.356422] env[61728]: DEBUG oslo_concurrency.lockutils [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "c14034b5-330b-4fb1-af31-d530e3231992-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.356602] env[61728]: DEBUG oslo_concurrency.lockutils [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "c14034b5-330b-4fb1-af31-d530e3231992-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.356879] env[61728]: DEBUG oslo_concurrency.lockutils [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "c14034b5-330b-4fb1-af31-d530e3231992-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.358791] env[61728]: INFO nova.compute.manager [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Terminating instance [ 737.361637] env[61728]: DEBUG nova.compute.manager [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 737.361840] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 737.362689] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d5fc96e-a2ed-480f-a8ec-5927d866ade8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.367202] env[61728]: DEBUG nova.scheduler.client.report [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 737.378838] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 737.379716] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-13378397-d5c1-48e6-9bb1-66e0880c4dad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.398844] env[61728]: DEBUG nova.scheduler.client.report [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 737.399187] env[61728]: DEBUG nova.compute.provider_tree [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 737.418971] env[61728]: DEBUG nova.scheduler.client.report [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 737.429989] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464028, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.802365} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.429989] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] f2692e22-6c14-4bb3-a080-607f0731105d/f2692e22-6c14-4bb3-a080-607f0731105d.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 737.429989] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 737.429989] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9dc51f85-07c2-4611-8e2d-0d35d6149e8f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.439463] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 737.439463] env[61728]: value = "task-464036" [ 737.439463] env[61728]: _type = "Task" [ 737.439463] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.445513] env[61728]: DEBUG nova.scheduler.client.report [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 737.457385] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464036, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.459022] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 737.459288] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 737.459486] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleting the datastore file [datastore1] c14034b5-330b-4fb1-af31-d530e3231992 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 737.459903] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c5d64180-7acc-4c12-8057-e346acda8d10 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.468088] env[61728]: DEBUG oslo_vmware.api [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 737.468088] env[61728]: value = "task-464037" [ 737.468088] env[61728]: _type = "Task" [ 737.468088] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.478780] env[61728]: DEBUG oslo_vmware.api [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464037, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.602806] env[61728]: DEBUG nova.network.neutron [req-a0db059d-8928-457c-825d-b955e9deb2e3 req-71b04c35-0497-49a8-b598-dc17209901a9 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Updated VIF entry in instance network info cache for port 0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 737.603540] env[61728]: DEBUG nova.network.neutron [req-a0db059d-8928-457c-825d-b955e9deb2e3 req-71b04c35-0497-49a8-b598-dc17209901a9 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Updating instance_info_cache with network_info: [{"id": "0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3", "address": "fa:16:3e:f1:97:78", "network": {"id": "105779cb-8a8c-49ab-8ea4-00d3c7878a6b", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1826811954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f22617f7e1f4598b01ebb980c61a1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b046d2a-a7", "ovs_interfaceid": "0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.659795] env[61728]: DEBUG oslo_vmware.api [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-464031, 'name': ReconfigVM_Task, 'duration_secs': 0.200058} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.659795] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-121946', 'volume_id': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'name': 'volume-2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0f75629b-bdec-4593-aceb-cd0478972ccc', 'attached_at': '', 'detached_at': '', 'volume_id': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4', 'serial': '2eb2f0fc-4130-44f7-bddb-eb01dbadccc4'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 737.659795] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 737.661341] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e715b9-8ca0-43fa-ab18-b09b101e1f0e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.669650] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 737.672860] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-72eca508-9a47-43a1-8509-9b1f9213a317 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.740296] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464034, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.751537] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 737.751537] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 737.751537] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Deleting the datastore file [datastore1] 0f75629b-bdec-4593-aceb-cd0478972ccc {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 737.751537] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de1a7e97-b1fb-451c-acda-d18c7ae5e99e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.761200] env[61728]: DEBUG oslo_vmware.api [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Waiting for the task: (returnval){ [ 737.761200] env[61728]: value = "task-464039" [ 737.761200] env[61728]: _type = "Task" [ 737.761200] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.773752] env[61728]: DEBUG oslo_vmware.api [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-464039, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.951136] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464036, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078671} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.953951] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 737.954780] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26500c32-442c-44a2-821f-1dc058dd9214 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.979835] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] f2692e22-6c14-4bb3-a080-607f0731105d/f2692e22-6c14-4bb3-a080-607f0731105d.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 737.988848] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06b23033-3e65-4017-aa6b-ee51ebcf4a6f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.013024] env[61728]: DEBUG oslo_vmware.api [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464037, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.391194} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.013024] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 738.013024] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 738.013024] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 738.013024] env[61728]: INFO nova.compute.manager [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Took 0.65 seconds to destroy the instance on the hypervisor. [ 738.013024] env[61728]: DEBUG oslo.service.loopingcall [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 738.013024] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 738.013024] env[61728]: value = "task-464040" [ 738.013024] env[61728]: _type = "Task" [ 738.013024] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.014215] env[61728]: DEBUG nova.compute.manager [-] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 738.014215] env[61728]: DEBUG nova.network.neutron [-] [instance: c14034b5-330b-4fb1-af31-d530e3231992] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 738.029317] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464040, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.109219] env[61728]: DEBUG oslo_concurrency.lockutils [req-a0db059d-8928-457c-825d-b955e9deb2e3 req-71b04c35-0497-49a8-b598-dc17209901a9 service nova] Releasing lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.154542] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba18f9a-02c8-43cd-989b-c508f77bd8c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.164404] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f1cb2e-75f8-4538-a028-79c2b4000880 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.200308] env[61728]: DEBUG nova.network.neutron [req-5f0c99ef-6437-44b6-9cab-bc8ec644ecbb req-92ea3252-f91c-4acb-b98f-3fcfe43e00f2 service nova] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updated VIF entry in instance network info cache for port cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 738.200308] env[61728]: DEBUG nova.network.neutron [req-5f0c99ef-6437-44b6-9cab-bc8ec644ecbb req-92ea3252-f91c-4acb-b98f-3fcfe43e00f2 service nova] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating instance_info_cache with network_info: [{"id": "cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd", "address": "fa:16:3e:fc:46:55", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc2e7ae1-f4", "ovs_interfaceid": "cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.205812] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9010eeb5-8fc5-49bc-9a43-9a3e6349d9b2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.216035] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c6f39ce-3f76-4661-bb50-d81f0b311f5a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.232276] env[61728]: DEBUG nova.compute.provider_tree [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 738.242962] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464034, 'name': CreateVM_Task, 'duration_secs': 0.55315} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.244260] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 738.244260] env[61728]: DEBUG oslo_concurrency.lockutils [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.244260] env[61728]: DEBUG oslo_concurrency.lockutils [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.244498] env[61728]: DEBUG oslo_concurrency.lockutils [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 738.244757] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32037b15-5354-4916-b083-c5bd454e3f95 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.251854] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 738.251854] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb59ac-03c0-d267-514b-8c6deccc5ef6" [ 738.251854] env[61728]: _type = "Task" [ 738.251854] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.260887] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb59ac-03c0-d267-514b-8c6deccc5ef6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.274150] env[61728]: DEBUG oslo_vmware.api [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Task: {'id': task-464039, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.268215} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.274553] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 738.274696] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 738.274894] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 738.359237] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 738.359237] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-29c19a61-4509-426c-9dfc-1e3197da672b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.369446] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b086f114-7e5b-4e50-b8a6-cce49823370f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.415124] env[61728]: ERROR nova.compute.manager [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Failed to detach volume 2eb2f0fc-4130-44f7-bddb-eb01dbadccc4 from /dev/sda: nova.exception.InstanceNotFound: Instance 0f75629b-bdec-4593-aceb-cd0478972ccc could not be found. [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Traceback (most recent call last): [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] self.driver.rebuild(**kwargs) [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/virt/driver.py", line 390, in rebuild [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] raise NotImplementedError() [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] NotImplementedError [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] During handling of the above exception, another exception occurred: [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Traceback (most recent call last): [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] self.driver.detach_volume(context, old_connection_info, [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 552, in detach_volume [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] return self._volumeops.detach_volume(connection_info, instance) [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] self._detach_volume_vmdk(connection_info, instance) [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] stable_ref.fetch_moref(session) [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] raise exception.InstanceNotFound(instance_id=self._uuid) [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] nova.exception.InstanceNotFound: Instance 0f75629b-bdec-4593-aceb-cd0478972ccc could not be found. [ 738.415124] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] [ 738.537167] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464040, 'name': ReconfigVM_Task, 'duration_secs': 0.322405} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.537167] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Reconfigured VM instance instance-00000035 to attach disk [datastore1] f2692e22-6c14-4bb3-a080-607f0731105d/f2692e22-6c14-4bb3-a080-607f0731105d.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 738.537729] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-14c9d609-33e5-436a-a56f-11c178331298 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.546073] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 738.546073] env[61728]: value = "task-464041" [ 738.546073] env[61728]: _type = "Task" [ 738.546073] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.556168] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464041, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.592969] env[61728]: DEBUG nova.compute.utils [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Build of instance 0f75629b-bdec-4593-aceb-cd0478972ccc aborted: Failed to rebuild volume backed instance. {{(pid=61728) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 738.594429] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "06109957-2c3b-404e-a43e-dd34ece39096" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.594656] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "06109957-2c3b-404e-a43e-dd34ece39096" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.595538] env[61728]: ERROR nova.compute.manager [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 0f75629b-bdec-4593-aceb-cd0478972ccc aborted: Failed to rebuild volume backed instance. [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Traceback (most recent call last): [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] self.driver.rebuild(**kwargs) [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/virt/driver.py", line 390, in rebuild [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] raise NotImplementedError() [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] NotImplementedError [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] During handling of the above exception, another exception occurred: [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Traceback (most recent call last): [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] self._detach_root_volume(context, instance, root_bdm) [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] with excutils.save_and_reraise_exception(): [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] self.force_reraise() [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] raise self.value [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] self.driver.detach_volume(context, old_connection_info, [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 552, in detach_volume [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] return self._volumeops.detach_volume(connection_info, instance) [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] self._detach_volume_vmdk(connection_info, instance) [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] stable_ref.fetch_moref(session) [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] raise exception.InstanceNotFound(instance_id=self._uuid) [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] nova.exception.InstanceNotFound: Instance 0f75629b-bdec-4593-aceb-cd0478972ccc could not be found. [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] During handling of the above exception, another exception occurred: [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Traceback (most recent call last): [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] yield [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 738.595538] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] self._do_rebuild_instance_with_claim( [ 738.596866] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 738.596866] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] self._do_rebuild_instance( [ 738.596866] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 738.596866] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] self._rebuild_default_impl(**kwargs) [ 738.596866] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 738.596866] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] self._rebuild_volume_backed_instance( [ 738.596866] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 738.596866] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] raise exception.BuildAbortException( [ 738.596866] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] nova.exception.BuildAbortException: Build of instance 0f75629b-bdec-4593-aceb-cd0478972ccc aborted: Failed to rebuild volume backed instance. [ 738.596866] env[61728]: ERROR nova.compute.manager [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] [ 738.710377] env[61728]: DEBUG oslo_concurrency.lockutils [req-5f0c99ef-6437-44b6-9cab-bc8ec644ecbb req-92ea3252-f91c-4acb-b98f-3fcfe43e00f2 service nova] Releasing lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.759507] env[61728]: ERROR nova.scheduler.client.report [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [req-5b3082a6-3b25-4660-99b2-952455bc9377] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5b3082a6-3b25-4660-99b2-952455bc9377"}]} [ 738.770013] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb59ac-03c0-d267-514b-8c6deccc5ef6, 'name': SearchDatastore_Task, 'duration_secs': 0.025034} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.770134] env[61728]: DEBUG oslo_concurrency.lockutils [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.770520] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 738.770659] env[61728]: DEBUG oslo_concurrency.lockutils [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.771233] env[61728]: DEBUG oslo_concurrency.lockutils [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.771233] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 738.771405] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8820f0b1-858c-4fa6-8f51-9ae2ea255e7a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.779166] env[61728]: DEBUG nova.scheduler.client.report [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 738.785523] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 738.785619] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 738.786587] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1344d341-20d8-4ead-9833-48c46f84c5ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.792014] env[61728]: DEBUG nova.scheduler.client.report [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 738.792319] env[61728]: DEBUG nova.compute.provider_tree [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 738.795647] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 738.795647] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a71fcf-19fc-3801-215d-dad66cc4ca61" [ 738.795647] env[61728]: _type = "Task" [ 738.795647] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.804375] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a71fcf-19fc-3801-215d-dad66cc4ca61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.805454] env[61728]: DEBUG nova.scheduler.client.report [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 738.825161] env[61728]: DEBUG nova.scheduler.client.report [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 738.865325] env[61728]: DEBUG nova.network.neutron [-] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.059539] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464041, 'name': Rename_Task, 'duration_secs': 0.153026} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.059873] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 739.060140] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e79a3a24-8389-477f-9d8e-a74bf088477a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.067536] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 739.067536] env[61728]: value = "task-464042" [ 739.067536] env[61728]: _type = "Task" [ 739.067536] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.080969] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464042, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.206961] env[61728]: DEBUG oslo_concurrency.lockutils [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Acquiring lock "8c4b7227-1a01-4400-beb8-2df27e17e329" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.207172] env[61728]: DEBUG oslo_concurrency.lockutils [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Lock "8c4b7227-1a01-4400-beb8-2df27e17e329" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.310135] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a71fcf-19fc-3801-215d-dad66cc4ca61, 'name': SearchDatastore_Task, 'duration_secs': 0.016936} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.311236] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b097ad8-f82b-48a8-ba1d-0b30930e6503 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.319154] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 739.319154] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524abb71-20d9-4083-76e8-69878389d3dd" [ 739.319154] env[61728]: _type = "Task" [ 739.319154] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.328743] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524abb71-20d9-4083-76e8-69878389d3dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.372132] env[61728]: INFO nova.compute.manager [-] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Took 1.36 seconds to deallocate network for instance. [ 739.393063] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29450735-4110-4ae3-a5f9-df1c05cd6c2c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.401789] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a98903d-5792-4520-ad0e-3884787cd124 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.436448] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0704e65b-caf6-4195-89bc-77f3ad57c82a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.446977] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ea7b65-0b0f-4c5a-8d64-5973ec875651 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.465615] env[61728]: DEBUG nova.compute.provider_tree [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 739.495272] env[61728]: DEBUG nova.compute.manager [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Received event network-changed-0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 739.495483] env[61728]: DEBUG nova.compute.manager [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Refreshing instance network info cache due to event network-changed-0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 739.495708] env[61728]: DEBUG oslo_concurrency.lockutils [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] Acquiring lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.495859] env[61728]: DEBUG oslo_concurrency.lockutils [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] Acquired lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.496033] env[61728]: DEBUG nova.network.neutron [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Refreshing network info cache for port 0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 739.579830] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464042, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.628525] env[61728]: DEBUG nova.compute.manager [req-b2ad6242-8511-4023-9cc5-bf1334a0015c req-9032299b-5aa1-4a62-9993-2766418edea2 service nova] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Received event network-vif-deleted-fa8afb65-af92-44ff-932f-b6e166dc8609 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 739.712674] env[61728]: INFO nova.compute.manager [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Detaching volume bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7 [ 739.752213] env[61728]: INFO nova.virt.block_device [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Attempting to driver detach volume bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7 from mountpoint /dev/sdb [ 739.752461] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 739.752658] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122074', 'volume_id': 'bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7', 'name': 'volume-bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8c4b7227-1a01-4400-beb8-2df27e17e329', 'attached_at': '', 'detached_at': '', 'volume_id': 'bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7', 'serial': 'bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 739.753627] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-807139c6-9eaf-46b2-8850-c31c7316d022 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.776146] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54faf1f6-e179-453f-988f-7cb60f613a00 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.786019] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d05eef-c1a5-4810-beb2-7d675fd318d9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.806742] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074ace2b-3417-43a3-a414-8e80c73d2867 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.823665] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] The volume has not been displaced from its original location: [datastore1] volume-bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7/volume-bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 739.829284] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Reconfiguring VM instance instance-0000000f to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 739.833945] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fddae6b8-0699-4cc0-b330-d9f44c2ac2ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.856240] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524abb71-20d9-4083-76e8-69878389d3dd, 'name': SearchDatastore_Task, 'duration_secs': 0.015398} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.857986] env[61728]: DEBUG oslo_concurrency.lockutils [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.858482] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] dc8a587f-3dc8-43b3-b986-2096e3c4e1f1/dc8a587f-3dc8-43b3-b986-2096e3c4e1f1.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 739.858988] env[61728]: DEBUG oslo_vmware.api [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Waiting for the task: (returnval){ [ 739.858988] env[61728]: value = "task-464043" [ 739.858988] env[61728]: _type = "Task" [ 739.858988] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.859458] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5e41f60-c045-4899-94c2-7316f07fa691 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.872335] env[61728]: DEBUG oslo_vmware.api [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Task: {'id': task-464043, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.874057] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 739.874057] env[61728]: value = "task-464044" [ 739.874057] env[61728]: _type = "Task" [ 739.874057] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.880142] env[61728]: DEBUG oslo_concurrency.lockutils [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.884128] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464044, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.004294] env[61728]: DEBUG nova.scheduler.client.report [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 93 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 740.004987] env[61728]: DEBUG nova.compute.provider_tree [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 93 to 94 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 740.005359] env[61728]: DEBUG nova.compute.provider_tree [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 740.081701] env[61728]: DEBUG oslo_vmware.api [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464042, 'name': PowerOnVM_Task, 'duration_secs': 0.527529} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.081701] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 740.081959] env[61728]: INFO nova.compute.manager [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Took 8.26 seconds to spawn the instance on the hypervisor. [ 740.082642] env[61728]: DEBUG nova.compute.manager [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 740.083388] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c2ff88-f37e-4677-ba6e-e3b656549c6f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.258776] env[61728]: DEBUG nova.network.neutron [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Updated VIF entry in instance network info cache for port 0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 740.258776] env[61728]: DEBUG nova.network.neutron [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Updating instance_info_cache with network_info: [{"id": "0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3", "address": "fa:16:3e:f1:97:78", "network": {"id": "105779cb-8a8c-49ab-8ea4-00d3c7878a6b", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1826811954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f22617f7e1f4598b01ebb980c61a1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b046d2a-a7", "ovs_interfaceid": "0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.374935] env[61728]: DEBUG oslo_vmware.api [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Task: {'id': task-464043, 'name': ReconfigVM_Task, 'duration_secs': 0.311539} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.378325] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Reconfigured VM instance instance-0000000f to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 740.383292] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-170ba454-e374-4ab0-9779-72a2dafb6f0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.402412] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464044, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.404110] env[61728]: DEBUG oslo_vmware.api [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Waiting for the task: (returnval){ [ 740.404110] env[61728]: value = "task-464045" [ 740.404110] env[61728]: _type = "Task" [ 740.404110] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.416373] env[61728]: DEBUG oslo_vmware.api [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Task: {'id': task-464045, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.513648] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 5.465s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.513648] env[61728]: INFO nova.compute.manager [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Migrating [ 740.522646] env[61728]: DEBUG oslo_concurrency.lockutils [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.026s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.525986] env[61728]: INFO nova.compute.claims [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 740.606331] env[61728]: INFO nova.compute.manager [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Took 47.70 seconds to build instance. [ 740.614397] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.660405] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "2976d71f-7322-4640-b734-645dbb70e0c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.661125] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "2976d71f-7322-4640-b734-645dbb70e0c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.761401] env[61728]: DEBUG oslo_concurrency.lockutils [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] Releasing lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.761401] env[61728]: DEBUG nova.compute.manager [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Received event network-changed-f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 740.761401] env[61728]: DEBUG nova.compute.manager [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Refreshing instance network info cache due to event network-changed-f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 740.761401] env[61728]: DEBUG oslo_concurrency.lockutils [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] Acquiring lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.761401] env[61728]: DEBUG oslo_concurrency.lockutils [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] Acquired lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.761401] env[61728]: DEBUG nova.network.neutron [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Refreshing network info cache for port f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 740.885963] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464044, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.674838} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.886299] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] dc8a587f-3dc8-43b3-b986-2096e3c4e1f1/dc8a587f-3dc8-43b3-b986-2096e3c4e1f1.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 740.886528] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 740.886790] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-79d990d5-b69c-4ed8-85fa-61cfbeae6417 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.897408] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 740.897408] env[61728]: value = "task-464046" [ 740.897408] env[61728]: _type = "Task" [ 740.897408] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.914138] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464046, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.924913] env[61728]: DEBUG oslo_vmware.api [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Task: {'id': task-464045, 'name': ReconfigVM_Task, 'duration_secs': 0.321652} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.925283] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122074', 'volume_id': 'bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7', 'name': 'volume-bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8c4b7227-1a01-4400-beb8-2df27e17e329', 'attached_at': '', 'detached_at': '', 'volume_id': 'bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7', 'serial': 'bdbfa2fe-cd5c-4f14-b52d-3ed995b50ba7'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 741.044027] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.044027] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.044027] env[61728]: DEBUG nova.network.neutron [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.108460] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8ef975e-1f79-42a6-8838-df4c4811fb55 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "f2692e22-6c14-4bb3-a080-607f0731105d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.966s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.407995] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464046, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.10252} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.408307] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 741.409100] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a6a65b-d5bc-4d74-a9d4-361081563442 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.430894] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] dc8a587f-3dc8-43b3-b986-2096e3c4e1f1/dc8a587f-3dc8-43b3-b986-2096e3c4e1f1.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 741.431686] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e0aa544-a145-4e13-9576-fa745cd8b6eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.461261] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 741.461261] env[61728]: value = "task-464047" [ 741.461261] env[61728]: _type = "Task" [ 741.461261] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.472195] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464047, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.501157] env[61728]: DEBUG nova.objects.instance [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Lazy-loading 'flavor' on Instance uuid 8c4b7227-1a01-4400-beb8-2df27e17e329 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 741.533460] env[61728]: DEBUG nova.network.neutron [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Updated VIF entry in instance network info cache for port f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 741.533460] env[61728]: DEBUG nova.network.neutron [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Updating instance_info_cache with network_info: [{"id": "f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1", "address": "fa:16:3e:f7:31:b7", "network": {"id": "105779cb-8a8c-49ab-8ea4-00d3c7878a6b", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1826811954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f22617f7e1f4598b01ebb980c61a1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9bdd9c4-3e", "ovs_interfaceid": "f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.611580] env[61728]: DEBUG nova.compute.manager [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 741.752579] env[61728]: DEBUG nova.network.neutron [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance_info_cache with network_info: [{"id": "4d70757b-d260-4aa3-8914-a38eb567f7b8", "address": "fa:16:3e:2a:f5:67", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.112", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d70757b-d2", "ovs_interfaceid": "4d70757b-d260-4aa3-8914-a38eb567f7b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.805402] env[61728]: DEBUG nova.compute.manager [req-9fad9f60-b309-4244-9423-b34360b53758 req-511a69d3-0a2b-4bc7-b012-3e954761e201 service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Received event network-changed-f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 741.805781] env[61728]: DEBUG nova.compute.manager [req-9fad9f60-b309-4244-9423-b34360b53758 req-511a69d3-0a2b-4bc7-b012-3e954761e201 service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Refreshing instance network info cache due to event network-changed-f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 741.806129] env[61728]: DEBUG oslo_concurrency.lockutils [req-9fad9f60-b309-4244-9423-b34360b53758 req-511a69d3-0a2b-4bc7-b012-3e954761e201 service nova] Acquiring lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.881371] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Acquiring lock "0f75629b-bdec-4593-aceb-cd0478972ccc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.881528] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Lock "0f75629b-bdec-4593-aceb-cd0478972ccc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.881744] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Acquiring lock "0f75629b-bdec-4593-aceb-cd0478972ccc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.881956] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Lock "0f75629b-bdec-4593-aceb-cd0478972ccc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.882124] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Lock "0f75629b-bdec-4593-aceb-cd0478972ccc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.907009] env[61728]: INFO nova.compute.manager [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Terminating instance [ 741.909746] env[61728]: DEBUG nova.compute.manager [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 741.910470] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a9173ea3-fc8c-40c7-872d-8a1d20eaddeb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.924022] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49266cf-266b-45ea-b91b-76e8ab3dac6f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.966177] env[61728]: DEBUG oslo_concurrency.lockutils [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.966553] env[61728]: DEBUG oslo_concurrency.lockutils [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.966673] env[61728]: DEBUG oslo_concurrency.lockutils [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.966851] env[61728]: DEBUG oslo_concurrency.lockutils [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.967021] env[61728]: DEBUG oslo_concurrency.lockutils [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.969146] env[61728]: WARNING nova.virt.vmwareapi.driver [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 0f75629b-bdec-4593-aceb-cd0478972ccc could not be found. [ 741.969278] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 741.973101] env[61728]: INFO nova.compute.manager [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Terminating instance [ 741.976774] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64c59d33-465c-4b98-833d-fbcba7c65602 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.979444] env[61728]: DEBUG nova.compute.manager [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 741.979676] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 741.980474] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744fa549-8752-4cd7-bc71-437514047112 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.989794] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464047, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.993811] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 741.995816] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-07f61d07-acdb-4ba7-a957-09000e19f47b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.002783] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c33cfb3-82bd-4fb1-8a48-da0de66e5c81 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.029408] env[61728]: DEBUG oslo_vmware.api [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 742.029408] env[61728]: value = "task-464048" [ 742.029408] env[61728]: _type = "Task" [ 742.029408] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.036449] env[61728]: DEBUG oslo_concurrency.lockutils [req-5b95762e-6d44-4c90-ae45-00c202899ba2 req-f6838caa-a061-4880-b407-9c4837ddf49c service nova] Releasing lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.039258] env[61728]: DEBUG oslo_concurrency.lockutils [req-9fad9f60-b309-4244-9423-b34360b53758 req-511a69d3-0a2b-4bc7-b012-3e954761e201 service nova] Acquired lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.042059] env[61728]: DEBUG nova.network.neutron [req-9fad9f60-b309-4244-9423-b34360b53758 req-511a69d3-0a2b-4bc7-b012-3e954761e201 service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Refreshing network info cache for port f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 742.058822] env[61728]: WARNING nova.virt.vmwareapi.vmops [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0f75629b-bdec-4593-aceb-cd0478972ccc could not be found. [ 742.059043] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 742.059246] env[61728]: INFO nova.compute.manager [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Took 0.15 seconds to destroy the instance on the hypervisor. [ 742.059531] env[61728]: DEBUG oslo.service.loopingcall [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.067868] env[61728]: DEBUG nova.compute.manager [-] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.067993] env[61728]: DEBUG nova.network.neutron [-] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 742.069625] env[61728]: DEBUG oslo_vmware.api [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464048, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.071807] env[61728]: DEBUG oslo_vmware.rw_handles [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524bf492-0faa-abc5-7153-3352deadaf5b/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 742.074593] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a1ead3d-5c4e-47b7-ac78-b7c6fb5688c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.081590] env[61728]: DEBUG oslo_vmware.rw_handles [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524bf492-0faa-abc5-7153-3352deadaf5b/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 742.081804] env[61728]: ERROR oslo_vmware.rw_handles [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524bf492-0faa-abc5-7153-3352deadaf5b/disk-0.vmdk due to incomplete transfer. [ 742.082113] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-24aa41a7-213e-4780-9a04-1164a27db053 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.107016] env[61728]: DEBUG oslo_vmware.rw_handles [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524bf492-0faa-abc5-7153-3352deadaf5b/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 742.107016] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Uploaded image 7ac64905-f9d8-4de1-beb4-cd6197d96f1e to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 742.108771] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 742.109103] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5f78d23c-9e78-4551-be94-258e3c52635f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.119163] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 742.119163] env[61728]: value = "task-464049" [ 742.119163] env[61728]: _type = "Task" [ 742.119163] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.133453] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464049, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.135800] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.255657] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Releasing lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.258968] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15034572-f717-4fbc-8583-62957ee2621d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.271168] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2693d231-79e0-49f3-91ab-539ffcf58a09 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.304108] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a725e910-6c33-4278-a40f-5e701b282dbd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.312771] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d68b4435-5f47-4d2a-b182-df9de372c95a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.336097] env[61728]: DEBUG nova.compute.provider_tree [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.474316] env[61728]: DEBUG nova.network.neutron [req-9fad9f60-b309-4244-9423-b34360b53758 req-511a69d3-0a2b-4bc7-b012-3e954761e201 service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Updated VIF entry in instance network info cache for port f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 742.474701] env[61728]: DEBUG nova.network.neutron [req-9fad9f60-b309-4244-9423-b34360b53758 req-511a69d3-0a2b-4bc7-b012-3e954761e201 service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Updating instance_info_cache with network_info: [{"id": "f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1", "address": "fa:16:3e:f7:31:b7", "network": {"id": "105779cb-8a8c-49ab-8ea4-00d3c7878a6b", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1826811954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f22617f7e1f4598b01ebb980c61a1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9bdd9c4-3e", "ovs_interfaceid": "f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.483235] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464047, 'name': ReconfigVM_Task, 'duration_secs': 0.579151} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.483576] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Reconfigured VM instance instance-00000036 to attach disk [datastore1] dc8a587f-3dc8-43b3-b986-2096e3c4e1f1/dc8a587f-3dc8-43b3-b986-2096e3c4e1f1.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 742.484211] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ecb5fbcb-15b0-43ac-b4c9-cd96a80af6d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.490994] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 742.490994] env[61728]: value = "task-464050" [ 742.490994] env[61728]: _type = "Task" [ 742.490994] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.509834] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464050, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.521244] env[61728]: DEBUG oslo_concurrency.lockutils [None req-804f67ac-e214-4f8f-8178-a508d16ca8f0 tempest-VolumesAssistedSnapshotsTest-416812987 tempest-VolumesAssistedSnapshotsTest-416812987-project-admin] Lock "8c4b7227-1a01-4400-beb8-2df27e17e329" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.314s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.544596] env[61728]: DEBUG oslo_vmware.api [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464048, 'name': PowerOffVM_Task, 'duration_secs': 0.203087} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.544596] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 742.544596] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 742.544596] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-14f8aca0-9348-4447-8dac-f25dc203a4db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.617744] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 742.617744] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 742.618291] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Deleting the datastore file [datastore1] 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 742.618608] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd5a4aca-880d-4b0d-8d99-fc6f03279d50 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.633155] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464049, 'name': Destroy_Task, 'duration_secs': 0.329058} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.633961] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Destroyed the VM [ 742.635104] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 742.635104] env[61728]: DEBUG oslo_vmware.api [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 742.635104] env[61728]: value = "task-464052" [ 742.635104] env[61728]: _type = "Task" [ 742.635104] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.635358] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9001228a-d429-48ec-bb04-dc99d80e2212 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.648411] env[61728]: DEBUG oslo_vmware.api [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464052, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.650491] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 742.650491] env[61728]: value = "task-464053" [ 742.650491] env[61728]: _type = "Task" [ 742.650491] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.663066] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464053, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.841165] env[61728]: DEBUG nova.scheduler.client.report [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.978288] env[61728]: DEBUG oslo_concurrency.lockutils [req-9fad9f60-b309-4244-9423-b34360b53758 req-511a69d3-0a2b-4bc7-b012-3e954761e201 service nova] Releasing lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.003979] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464050, 'name': Rename_Task, 'duration_secs': 0.166366} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.004348] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 743.004616] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90fef980-a3c9-47f5-a1ed-cf81468f46ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.011917] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 743.011917] env[61728]: value = "task-464055" [ 743.011917] env[61728]: _type = "Task" [ 743.011917] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.021439] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464055, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.094605] env[61728]: DEBUG nova.compute.manager [req-0f645efd-676c-4bf0-9bdd-6e35dc0ccb3d req-eae61e09-80e4-46be-881e-57621d74dae2 service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Received event network-vif-deleted-1915184e-ab7a-409c-a8d2-ccbdca5e942e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 743.094840] env[61728]: INFO nova.compute.manager [req-0f645efd-676c-4bf0-9bdd-6e35dc0ccb3d req-eae61e09-80e4-46be-881e-57621d74dae2 service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Neutron deleted interface 1915184e-ab7a-409c-a8d2-ccbdca5e942e; detaching it from the instance and deleting it from the info cache [ 743.095167] env[61728]: DEBUG nova.network.neutron [req-0f645efd-676c-4bf0-9bdd-6e35dc0ccb3d req-eae61e09-80e4-46be-881e-57621d74dae2 service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.148637] env[61728]: DEBUG oslo_vmware.api [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464052, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1849} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.149018] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 743.149124] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 743.149310] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 743.149487] env[61728]: INFO nova.compute.manager [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Took 1.17 seconds to destroy the instance on the hypervisor. [ 743.149768] env[61728]: DEBUG oslo.service.loopingcall [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.149974] env[61728]: DEBUG nova.compute.manager [-] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.150083] env[61728]: DEBUG nova.network.neutron [-] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 743.160191] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464053, 'name': RemoveSnapshot_Task} progress is 78%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.207521] env[61728]: DEBUG nova.network.neutron [-] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.348239] env[61728]: DEBUG oslo_concurrency.lockutils [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.826s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.348801] env[61728]: DEBUG nova.compute.manager [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 743.352029] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 41.875s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.352029] env[61728]: DEBUG nova.objects.instance [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61728) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 743.523627] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464055, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.598611] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c95c75f-6f85-4633-863c-791a129c5bfa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.610933] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-070bf793-22b8-461b-a5bb-60667b9550d4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.655254] env[61728]: DEBUG nova.compute.manager [req-0f645efd-676c-4bf0-9bdd-6e35dc0ccb3d req-eae61e09-80e4-46be-881e-57621d74dae2 service nova] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Detach interface failed, port_id=1915184e-ab7a-409c-a8d2-ccbdca5e942e, reason: Instance 0f75629b-bdec-4593-aceb-cd0478972ccc could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 743.664641] env[61728]: DEBUG oslo_vmware.api [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464053, 'name': RemoveSnapshot_Task, 'duration_secs': 0.85518} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.664926] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 743.665271] env[61728]: INFO nova.compute.manager [None req-419f3213-2b18-4511-b30e-730b9c59266f tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Took 18.43 seconds to snapshot the instance on the hypervisor. [ 743.709576] env[61728]: INFO nova.compute.manager [-] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Took 1.64 seconds to deallocate network for instance. [ 743.778894] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5753e28e-486f-44d9-a589-23e290641885 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.800363] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance '1d9114cb-a0af-4874-962f-27237b3c89cc' progress to 0 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 743.866445] env[61728]: DEBUG nova.compute.utils [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 743.866445] env[61728]: DEBUG nova.compute.manager [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Not allocating networking since 'none' was specified. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 743.946833] env[61728]: DEBUG nova.network.neutron [-] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.028232] env[61728]: DEBUG oslo_vmware.api [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464055, 'name': PowerOnVM_Task, 'duration_secs': 0.875048} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.028832] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 744.029124] env[61728]: INFO nova.compute.manager [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Took 6.94 seconds to spawn the instance on the hypervisor. [ 744.029327] env[61728]: DEBUG nova.compute.manager [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 744.030155] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f80cf993-b44f-464e-a214-83226f5462b4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.244556] env[61728]: DEBUG nova.compute.manager [req-4c3e2b10-4efd-4604-8ad2-7378cba0b702 req-a4e346c0-2a1d-4503-8142-85665bf9ba3f service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Received event network-changed-0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 744.244556] env[61728]: DEBUG nova.compute.manager [req-4c3e2b10-4efd-4604-8ad2-7378cba0b702 req-a4e346c0-2a1d-4503-8142-85665bf9ba3f service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Refreshing instance network info cache due to event network-changed-0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 744.244781] env[61728]: DEBUG oslo_concurrency.lockutils [req-4c3e2b10-4efd-4604-8ad2-7378cba0b702 req-a4e346c0-2a1d-4503-8142-85665bf9ba3f service nova] Acquiring lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.244781] env[61728]: DEBUG oslo_concurrency.lockutils [req-4c3e2b10-4efd-4604-8ad2-7378cba0b702 req-a4e346c0-2a1d-4503-8142-85665bf9ba3f service nova] Acquired lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.244925] env[61728]: DEBUG nova.network.neutron [req-4c3e2b10-4efd-4604-8ad2-7378cba0b702 req-a4e346c0-2a1d-4503-8142-85665bf9ba3f service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Refreshing network info cache for port 0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 744.274577] env[61728]: INFO nova.compute.manager [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Took 0.56 seconds to detach 1 volumes for instance. [ 744.280882] env[61728]: DEBUG nova.compute.manager [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Deleting volume: 2eb2f0fc-4130-44f7-bddb-eb01dbadccc4 {{(pid=61728) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 744.307169] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 744.307493] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6403713f-5c02-4ce4-92e7-fae95d5d3903 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.318644] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 744.318644] env[61728]: value = "task-464056" [ 744.318644] env[61728]: _type = "Task" [ 744.318644] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.333749] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464056, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.366653] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1a7fa267-cb91-4ff0-856a-52825c9a2a75 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.370265] env[61728]: DEBUG nova.compute.manager [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 744.376471] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 41.938s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.376471] env[61728]: INFO nova.compute.claims [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.449837] env[61728]: INFO nova.compute.manager [-] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Took 1.30 seconds to deallocate network for instance. [ 744.548352] env[61728]: INFO nova.compute.manager [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Took 49.02 seconds to build instance. [ 744.829730] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.843840] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464056, 'name': PowerOffVM_Task, 'duration_secs': 0.243831} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.843840] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 744.843840] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance '1d9114cb-a0af-4874-962f-27237b3c89cc' progress to 17 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 744.960181] env[61728]: DEBUG oslo_concurrency.lockutils [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.986903] env[61728]: DEBUG nova.network.neutron [req-4c3e2b10-4efd-4604-8ad2-7378cba0b702 req-a4e346c0-2a1d-4503-8142-85665bf9ba3f service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Updated VIF entry in instance network info cache for port 0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 744.987342] env[61728]: DEBUG nova.network.neutron [req-4c3e2b10-4efd-4604-8ad2-7378cba0b702 req-a4e346c0-2a1d-4503-8142-85665bf9ba3f service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Updating instance_info_cache with network_info: [{"id": "0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3", "address": "fa:16:3e:f1:97:78", "network": {"id": "105779cb-8a8c-49ab-8ea4-00d3c7878a6b", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1826811954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f22617f7e1f4598b01ebb980c61a1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b046d2a-a7", "ovs_interfaceid": "0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.998539] env[61728]: DEBUG nova.compute.manager [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Stashing vm_state: active {{(pid=61728) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 745.052109] env[61728]: DEBUG oslo_concurrency.lockutils [None req-264c856c-a296-460b-b3c5-96a0bf515d9a tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "dc8a587f-3dc8-43b3-b986-2096e3c4e1f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.006s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.217909] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.236112] env[61728]: DEBUG nova.compute.manager [req-92bbc7bf-ddde-49af-b410-b04e6bef8f6e req-7cef70f9-487c-4ea7-906b-39cce3592afc service nova] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Received event network-vif-deleted-f9bdd9c4-3e34-40f0-94ca-c5103ae2b8d1 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 745.349154] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 745.349422] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 745.349634] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 745.349822] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 745.349977] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 745.350151] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 745.350366] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 745.350530] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 745.350711] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 745.350875] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 745.352975] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.358342] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b74d495-3472-48ee-9536-f1a02b97ef7a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.381234] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 745.381234] env[61728]: value = "task-464058" [ 745.381234] env[61728]: _type = "Task" [ 745.381234] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.385813] env[61728]: DEBUG nova.compute.manager [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 745.401225] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464058, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.421062] env[61728]: DEBUG nova.virt.hardware [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 745.421351] env[61728]: DEBUG nova.virt.hardware [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 745.421680] env[61728]: DEBUG nova.virt.hardware [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 745.421893] env[61728]: DEBUG nova.virt.hardware [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 745.422064] env[61728]: DEBUG nova.virt.hardware [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 745.422227] env[61728]: DEBUG nova.virt.hardware [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 745.422441] env[61728]: DEBUG nova.virt.hardware [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 745.422607] env[61728]: DEBUG nova.virt.hardware [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 745.422781] env[61728]: DEBUG nova.virt.hardware [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 745.422962] env[61728]: DEBUG nova.virt.hardware [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 745.423163] env[61728]: DEBUG nova.virt.hardware [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.424087] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8b8249-2606-4f1b-b034-da0b2f349fe8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.436081] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ca2eb7-3fa7-4fab-9f16-d2833eddabe9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.454904] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Instance VIF info [] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 745.462065] env[61728]: DEBUG oslo.service.loopingcall [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.466235] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e83145a-2481-411d-b037-a305a209074e] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 745.467625] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba1d7702-fd0e-47a1-ba50-60d6d608f003 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.491015] env[61728]: DEBUG oslo_concurrency.lockutils [req-4c3e2b10-4efd-4604-8ad2-7378cba0b702 req-a4e346c0-2a1d-4503-8142-85665bf9ba3f service nova] Releasing lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.492329] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 745.492329] env[61728]: value = "task-464059" [ 745.492329] env[61728]: _type = "Task" [ 745.492329] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.510138] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464059, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.532737] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.555223] env[61728]: DEBUG nova.compute.manager [None req-49c8b81b-62cf-43b4-9473-273d39065e09 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: fb962a28-c923-441f-9564-814c6bcb7cab] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 745.898350] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464058, 'name': ReconfigVM_Task, 'duration_secs': 0.264632} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.902044] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance '1d9114cb-a0af-4874-962f-27237b3c89cc' progress to 33 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 746.005354] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464059, 'name': CreateVM_Task, 'duration_secs': 0.482825} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.005354] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e83145a-2481-411d-b037-a305a209074e] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 746.005354] env[61728]: DEBUG oslo_concurrency.lockutils [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.005354] env[61728]: DEBUG oslo_concurrency.lockutils [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.005354] env[61728]: DEBUG oslo_concurrency.lockutils [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 746.005998] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a8cbaa8-01c0-4e7f-9259-4a8318d90780 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.016148] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 746.016148] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dd75f6-df53-fe28-d8d7-0c4987ac79e3" [ 746.016148] env[61728]: _type = "Task" [ 746.016148] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.027035] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dd75f6-df53-fe28-d8d7-0c4987ac79e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.057580] env[61728]: DEBUG nova.compute.manager [None req-49c8b81b-62cf-43b4-9473-273d39065e09 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: fb962a28-c923-441f-9564-814c6bcb7cab] Instance disappeared before build. {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 746.138654] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cc9945-66b4-460a-a87a-5de2f461db10 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.149977] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee236c74-a78d-4697-ad2f-fbcbfa905e67 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.186354] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f6370a-2da4-4776-9d97-118cd8ca05cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.198429] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0cdf1fb-e2f4-4587-b401-41bfc204a166 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.215319] env[61728]: DEBUG nova.compute.provider_tree [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.333243] env[61728]: DEBUG nova.compute.manager [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 746.334201] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c131ee-29e3-432e-be3f-8c11c72596e5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.348665] env[61728]: DEBUG nova.compute.manager [req-9ba3f384-adcd-4ed9-b3f7-df5123fc8253 req-afcf2a08-3f5e-4210-9509-df3462cb01d4 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Received event network-changed-0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 746.348904] env[61728]: DEBUG nova.compute.manager [req-9ba3f384-adcd-4ed9-b3f7-df5123fc8253 req-afcf2a08-3f5e-4210-9509-df3462cb01d4 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Refreshing instance network info cache due to event network-changed-0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 746.349146] env[61728]: DEBUG oslo_concurrency.lockutils [req-9ba3f384-adcd-4ed9-b3f7-df5123fc8253 req-afcf2a08-3f5e-4210-9509-df3462cb01d4 service nova] Acquiring lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.349318] env[61728]: DEBUG oslo_concurrency.lockutils [req-9ba3f384-adcd-4ed9-b3f7-df5123fc8253 req-afcf2a08-3f5e-4210-9509-df3462cb01d4 service nova] Acquired lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.349461] env[61728]: DEBUG nova.network.neutron [req-9ba3f384-adcd-4ed9-b3f7-df5123fc8253 req-afcf2a08-3f5e-4210-9509-df3462cb01d4 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Refreshing network info cache for port 0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 746.411122] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:24:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='ec856338-a71e-43d6-aa12-96f76d31f40f',id=36,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1304387059',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.411406] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.411645] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.411892] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.412108] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.412343] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.412600] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.412804] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.413945] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.413945] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.413945] env[61728]: DEBUG nova.virt.hardware [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.420385] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Reconfiguring VM instance instance-0000002b to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 746.420385] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0a98f00-45b2-4bfa-85ca-cfdcb9472ad5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.443519] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 746.443519] env[61728]: value = "task-464060" [ 746.443519] env[61728]: _type = "Task" [ 746.443519] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.452950] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464060, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.528201] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dd75f6-df53-fe28-d8d7-0c4987ac79e3, 'name': SearchDatastore_Task, 'duration_secs': 0.022172} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.528550] env[61728]: DEBUG oslo_concurrency.lockutils [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.528820] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 746.529080] env[61728]: DEBUG oslo_concurrency.lockutils [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.529255] env[61728]: DEBUG oslo_concurrency.lockutils [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.530030] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 746.530030] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fdba1e36-3200-40b2-a496-532bbb48b57f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.540418] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 746.540803] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 746.542093] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a08ffe7b-e7d4-4cea-b723-fad5212bba50 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.549321] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 746.549321] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bd10ef-cf32-e1bb-f33f-3c88f23af74d" [ 746.549321] env[61728]: _type = "Task" [ 746.549321] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.557904] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bd10ef-cf32-e1bb-f33f-3c88f23af74d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.572296] env[61728]: DEBUG oslo_concurrency.lockutils [None req-49c8b81b-62cf-43b4-9473-273d39065e09 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "fb962a28-c923-441f-9564-814c6bcb7cab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 9.408s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.693270] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Acquiring lock "8c4b7227-1a01-4400-beb8-2df27e17e329" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.693563] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Lock "8c4b7227-1a01-4400-beb8-2df27e17e329" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.693799] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Acquiring lock "8c4b7227-1a01-4400-beb8-2df27e17e329-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.694036] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Lock "8c4b7227-1a01-4400-beb8-2df27e17e329-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.694232] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Lock "8c4b7227-1a01-4400-beb8-2df27e17e329-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.696468] env[61728]: INFO nova.compute.manager [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Terminating instance [ 746.698439] env[61728]: DEBUG nova.compute.manager [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 746.698641] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 746.699482] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8729e18-c693-4d50-845e-c35d501b3dbc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.708409] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 746.708890] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d563fa9-a502-4cd8-9b73-e8e6db58d7ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.718017] env[61728]: DEBUG oslo_vmware.api [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Waiting for the task: (returnval){ [ 746.718017] env[61728]: value = "task-464061" [ 746.718017] env[61728]: _type = "Task" [ 746.718017] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.720969] env[61728]: DEBUG nova.scheduler.client.report [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.732672] env[61728]: DEBUG oslo_vmware.api [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-464061, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.853563] env[61728]: INFO nova.compute.manager [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] instance snapshotting [ 746.860690] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d28e460-3a0c-43f8-bc0c-b4baa1445ec3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.890300] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70ebde7-2cc6-401f-8ccb-1676d7e72ccf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.955621] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464060, 'name': ReconfigVM_Task, 'duration_secs': 0.239315} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.955621] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Reconfigured VM instance instance-0000002b to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 746.956368] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79cd0b7c-ef1e-41fe-b6c3-28690b8a3b26 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.982406] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] 1d9114cb-a0af-4874-962f-27237b3c89cc/1d9114cb-a0af-4874-962f-27237b3c89cc.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 746.985195] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c728d83c-6def-4e65-b59c-fdb515e0e33f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.005406] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 747.005406] env[61728]: value = "task-464062" [ 747.005406] env[61728]: _type = "Task" [ 747.005406] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.014753] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464062, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.061246] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bd10ef-cf32-e1bb-f33f-3c88f23af74d, 'name': SearchDatastore_Task, 'duration_secs': 0.013986} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.062185] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9637ae4-ac1b-4418-9bf0-196fbd01d1b5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.070290] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 747.070290] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ff597d-3f9e-fd2b-b4b3-0f85b381871e" [ 747.070290] env[61728]: _type = "Task" [ 747.070290] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.075132] env[61728]: DEBUG nova.compute.manager [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 747.080989] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ff597d-3f9e-fd2b-b4b3-0f85b381871e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.153479] env[61728]: DEBUG nova.network.neutron [req-9ba3f384-adcd-4ed9-b3f7-df5123fc8253 req-afcf2a08-3f5e-4210-9509-df3462cb01d4 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Updated VIF entry in instance network info cache for port 0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 747.153843] env[61728]: DEBUG nova.network.neutron [req-9ba3f384-adcd-4ed9-b3f7-df5123fc8253 req-afcf2a08-3f5e-4210-9509-df3462cb01d4 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Updating instance_info_cache with network_info: [{"id": "0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3", "address": "fa:16:3e:f1:97:78", "network": {"id": "105779cb-8a8c-49ab-8ea4-00d3c7878a6b", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1826811954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f22617f7e1f4598b01ebb980c61a1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77aa121f-8fb6-42f3-aaea-43addfe449b2", "external-id": "nsx-vlan-transportzone-288", "segmentation_id": 288, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b046d2a-a7", "ovs_interfaceid": "0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.228329] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.855s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.229095] env[61728]: DEBUG nova.compute.manager [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 747.232307] env[61728]: DEBUG oslo_vmware.api [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-464061, 'name': PowerOffVM_Task, 'duration_secs': 0.273888} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.235772] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 38.589s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.235772] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 747.235772] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 747.235772] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-752749a3-b76e-42be-96d6-279382c8d60f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.317297] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 747.317564] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 747.317774] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Deleting the datastore file [datastore1] 8c4b7227-1a01-4400-beb8-2df27e17e329 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 747.318104] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8271d99e-c2ce-4e3a-be59-175981630753 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.326170] env[61728]: DEBUG oslo_vmware.api [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Waiting for the task: (returnval){ [ 747.326170] env[61728]: value = "task-464064" [ 747.326170] env[61728]: _type = "Task" [ 747.326170] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.335839] env[61728]: DEBUG oslo_vmware.api [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-464064, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.412475] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 747.412900] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2f1a266a-49eb-463f-94c8-090d5e898797 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.422791] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 747.422791] env[61728]: value = "task-464065" [ 747.422791] env[61728]: _type = "Task" [ 747.422791] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.434706] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464065, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.488347] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquiring lock "6d55bb01-7855-4d86-ac57-ecae15ea590f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.488347] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Lock "6d55bb01-7855-4d86-ac57-ecae15ea590f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.517589] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464062, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.581585] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ff597d-3f9e-fd2b-b4b3-0f85b381871e, 'name': SearchDatastore_Task, 'duration_secs': 0.011941} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.584261] env[61728]: DEBUG oslo_concurrency.lockutils [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.584695] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2e83145a-2481-411d-b037-a305a209074e/2e83145a-2481-411d-b037-a305a209074e.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 747.586962] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c4a7fdb1-92af-47c0-8c69-01c17f105544 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.594977] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 747.594977] env[61728]: value = "task-464066" [ 747.594977] env[61728]: _type = "Task" [ 747.594977] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.600774] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.604465] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464066, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.657119] env[61728]: DEBUG oslo_concurrency.lockutils [req-9ba3f384-adcd-4ed9-b3f7-df5123fc8253 req-afcf2a08-3f5e-4210-9509-df3462cb01d4 service nova] Releasing lock "refresh_cache-0fb1192e-99f1-4469-b196-60df7eab8185" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.734298] env[61728]: DEBUG nova.compute.utils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 747.737075] env[61728]: DEBUG nova.compute.manager [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 747.737075] env[61728]: DEBUG nova.network.neutron [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 747.742455] env[61728]: INFO nova.compute.claims [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 747.790318] env[61728]: DEBUG nova.policy [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54c30aeeea3a4c1b8ed34430361c2a4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a02bd852c564acd99e12fd17279f028', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 747.836761] env[61728]: DEBUG oslo_vmware.api [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Task: {'id': task-464064, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198832} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.837056] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 747.837254] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 747.837677] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 747.838060] env[61728]: INFO nova.compute.manager [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Took 1.14 seconds to destroy the instance on the hypervisor. [ 747.838060] env[61728]: DEBUG oslo.service.loopingcall [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 747.838299] env[61728]: DEBUG nova.compute.manager [-] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 747.838426] env[61728]: DEBUG nova.network.neutron [-] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 747.939529] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464065, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.017824] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464062, 'name': ReconfigVM_Task, 'duration_secs': 0.618526} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.018179] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Reconfigured VM instance instance-0000002b to attach disk [datastore1] 1d9114cb-a0af-4874-962f-27237b3c89cc/1d9114cb-a0af-4874-962f-27237b3c89cc.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 748.018430] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance '1d9114cb-a0af-4874-962f-27237b3c89cc' progress to 50 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 748.107620] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464066, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.252352] env[61728]: DEBUG nova.compute.manager [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 748.261035] env[61728]: INFO nova.compute.resource_tracker [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating resource usage from migration 7ca3d98f-8659-4e9a-bbb2-d0f80e820da2 [ 748.278353] env[61728]: DEBUG nova.network.neutron [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Successfully created port: 3f69dcc2-f3d8-4b50-b760-85f6849a6c24 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.436729] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464065, 'name': CreateSnapshot_Task, 'duration_secs': 0.91406} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.436997] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 748.438207] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20c1957d-1bf2-4073-9b9b-f37942b1cc0b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.531122] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea34ed1b-1581-46e3-bdb6-c8f50c4f0330 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.555474] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4411a045-1dde-47bf-8275-e6219ff1ed3f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.576512] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance '1d9114cb-a0af-4874-962f-27237b3c89cc' progress to 67 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 748.607971] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464066, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681311} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.608560] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2e83145a-2481-411d-b037-a305a209074e/2e83145a-2481-411d-b037-a305a209074e.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 748.608827] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 748.609919] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-87301ddb-1082-4f2c-847a-37d46a198630 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.617778] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 748.617778] env[61728]: value = "task-464067" [ 748.617778] env[61728]: _type = "Task" [ 748.617778] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.634013] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464067, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.640724] env[61728]: DEBUG nova.compute.manager [req-47f66dd3-9959-491d-8247-1f6d9dfac0ab req-19008283-7709-4528-806a-af419a750af5 service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Received event network-vif-deleted-f77e3955-f8e6-43aa-ad48-d8bc4d078422 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 748.640816] env[61728]: INFO nova.compute.manager [req-47f66dd3-9959-491d-8247-1f6d9dfac0ab req-19008283-7709-4528-806a-af419a750af5 service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Neutron deleted interface f77e3955-f8e6-43aa-ad48-d8bc4d078422; detaching it from the instance and deleting it from the info cache [ 748.641240] env[61728]: DEBUG nova.network.neutron [req-47f66dd3-9959-491d-8247-1f6d9dfac0ab req-19008283-7709-4528-806a-af419a750af5 service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.947647] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecbd3132-f13e-4ae5-b4fd-bd8b658397f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.955945] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0614cae9-873b-40ef-ac6b-c956ec1e9865 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.995021] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 748.995580] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-42930ea5-d407-487e-8374-76a2038f49c3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.999034] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e8de87-a6dd-497b-b5b8-502232fc40db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.009625] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1bdd2d4-0578-4bdf-994d-f43915549f76 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.014364] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 749.014364] env[61728]: value = "task-464068" [ 749.014364] env[61728]: _type = "Task" [ 749.014364] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.027985] env[61728]: DEBUG nova.compute.provider_tree [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.035185] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464068, 'name': CloneVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.035185] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464068, 'name': CloneVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.084315] env[61728]: DEBUG nova.network.neutron [-] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.133291] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464067, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078592} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.133606] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 749.136141] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cbc7370-0a96-4d94-b1f1-5316911a8255 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.139626] env[61728]: DEBUG nova.network.neutron [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Port 4d70757b-d260-4aa3-8914-a38eb567f7b8 binding to destination host cpu-1 is already ACTIVE {{(pid=61728) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 749.162523] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 2e83145a-2481-411d-b037-a305a209074e/2e83145a-2481-411d-b037-a305a209074e.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 749.162523] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67f38e57-8cde-4e2e-8156-6bd4a3cd9e13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.165204] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-802a9785-6cfd-4341-ad23-bbc913d78b68 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.194111] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624a6716-7241-4fe3-9677-ea00ba36283c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.205054] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 749.205054] env[61728]: value = "task-464069" [ 749.205054] env[61728]: _type = "Task" [ 749.205054] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.217655] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464069, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.247028] env[61728]: DEBUG nova.compute.manager [req-47f66dd3-9959-491d-8247-1f6d9dfac0ab req-19008283-7709-4528-806a-af419a750af5 service nova] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Detach interface failed, port_id=f77e3955-f8e6-43aa-ad48-d8bc4d078422, reason: Instance 8c4b7227-1a01-4400-beb8-2df27e17e329 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 749.271748] env[61728]: DEBUG nova.compute.manager [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 749.301643] env[61728]: DEBUG nova.virt.hardware [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 749.301892] env[61728]: DEBUG nova.virt.hardware [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 749.302084] env[61728]: DEBUG nova.virt.hardware [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.302311] env[61728]: DEBUG nova.virt.hardware [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 749.302489] env[61728]: DEBUG nova.virt.hardware [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.302622] env[61728]: DEBUG nova.virt.hardware [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 749.302835] env[61728]: DEBUG nova.virt.hardware [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 749.303011] env[61728]: DEBUG nova.virt.hardware [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 749.303224] env[61728]: DEBUG nova.virt.hardware [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 749.303382] env[61728]: DEBUG nova.virt.hardware [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 749.303594] env[61728]: DEBUG nova.virt.hardware [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.304626] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf1e14c-f5fb-4269-9ddd-2c1a6d2f8250 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.314410] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d2035f2-83a4-41ee-bc38-59ca36dfc9c9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.525545] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464068, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.531699] env[61728]: DEBUG nova.scheduler.client.report [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.590100] env[61728]: INFO nova.compute.manager [-] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Took 1.75 seconds to deallocate network for instance. [ 749.717364] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464069, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.919731] env[61728]: DEBUG nova.network.neutron [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Successfully updated port: 3f69dcc2-f3d8-4b50-b760-85f6849a6c24 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 750.025987] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464068, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.037110] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.804s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.037388] env[61728]: INFO nova.compute.manager [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Migrating [ 750.044917] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.950s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.045269] env[61728]: DEBUG nova.objects.instance [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lazy-loading 'resources' on Instance uuid ca8d41dc-5719-4ade-b82a-b1d7f1c8def1 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 750.096984] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.162058] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "1d9114cb-a0af-4874-962f-27237b3c89cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.162305] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "1d9114cb-a0af-4874-962f-27237b3c89cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.162486] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "1d9114cb-a0af-4874-962f-27237b3c89cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.218444] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464069, 'name': ReconfigVM_Task, 'duration_secs': 0.865129} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.218734] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 2e83145a-2481-411d-b037-a305a209074e/2e83145a-2481-411d-b037-a305a209074e.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 750.219373] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e7ec5f0b-083a-42d3-9853-962bdbf68746 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.226655] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 750.226655] env[61728]: value = "task-464070" [ 750.226655] env[61728]: _type = "Task" [ 750.226655] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.235354] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464070, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.424084] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "refresh_cache-40ef29df-12d6-4274-846a-4a2c32752b8b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.424084] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "refresh_cache-40ef29df-12d6-4274-846a-4a2c32752b8b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.424084] env[61728]: DEBUG nova.network.neutron [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 750.527196] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464068, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.555048] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.555567] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.555567] env[61728]: DEBUG nova.network.neutron [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 750.738235] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464070, 'name': Rename_Task, 'duration_secs': 0.24085} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.741134] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 750.741665] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31dda465-2d7a-4258-b1c9-7ca57d803cdc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.750174] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 750.750174] env[61728]: value = "task-464071" [ 750.750174] env[61728]: _type = "Task" [ 750.750174] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.752375] env[61728]: DEBUG nova.compute.manager [req-e4e3d336-6ec3-40b6-98b5-35aae39ea9a7 req-203434d6-cca7-4558-bf5d-22e6494d8dc9 service nova] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Received event network-vif-plugged-3f69dcc2-f3d8-4b50-b760-85f6849a6c24 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 750.752590] env[61728]: DEBUG oslo_concurrency.lockutils [req-e4e3d336-6ec3-40b6-98b5-35aae39ea9a7 req-203434d6-cca7-4558-bf5d-22e6494d8dc9 service nova] Acquiring lock "40ef29df-12d6-4274-846a-4a2c32752b8b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.752804] env[61728]: DEBUG oslo_concurrency.lockutils [req-e4e3d336-6ec3-40b6-98b5-35aae39ea9a7 req-203434d6-cca7-4558-bf5d-22e6494d8dc9 service nova] Lock "40ef29df-12d6-4274-846a-4a2c32752b8b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.752973] env[61728]: DEBUG oslo_concurrency.lockutils [req-e4e3d336-6ec3-40b6-98b5-35aae39ea9a7 req-203434d6-cca7-4558-bf5d-22e6494d8dc9 service nova] Lock "40ef29df-12d6-4274-846a-4a2c32752b8b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.753155] env[61728]: DEBUG nova.compute.manager [req-e4e3d336-6ec3-40b6-98b5-35aae39ea9a7 req-203434d6-cca7-4558-bf5d-22e6494d8dc9 service nova] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] No waiting events found dispatching network-vif-plugged-3f69dcc2-f3d8-4b50-b760-85f6849a6c24 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 750.753322] env[61728]: WARNING nova.compute.manager [req-e4e3d336-6ec3-40b6-98b5-35aae39ea9a7 req-203434d6-cca7-4558-bf5d-22e6494d8dc9 service nova] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Received unexpected event network-vif-plugged-3f69dcc2-f3d8-4b50-b760-85f6849a6c24 for instance with vm_state building and task_state spawning. [ 750.753482] env[61728]: DEBUG nova.compute.manager [req-e4e3d336-6ec3-40b6-98b5-35aae39ea9a7 req-203434d6-cca7-4558-bf5d-22e6494d8dc9 service nova] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Received event network-changed-3f69dcc2-f3d8-4b50-b760-85f6849a6c24 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 750.753636] env[61728]: DEBUG nova.compute.manager [req-e4e3d336-6ec3-40b6-98b5-35aae39ea9a7 req-203434d6-cca7-4558-bf5d-22e6494d8dc9 service nova] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Refreshing instance network info cache due to event network-changed-3f69dcc2-f3d8-4b50-b760-85f6849a6c24. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 750.753808] env[61728]: DEBUG oslo_concurrency.lockutils [req-e4e3d336-6ec3-40b6-98b5-35aae39ea9a7 req-203434d6-cca7-4558-bf5d-22e6494d8dc9 service nova] Acquiring lock "refresh_cache-40ef29df-12d6-4274-846a-4a2c32752b8b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.765934] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464071, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.961177] env[61728]: DEBUG nova.network.neutron [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 751.030209] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464068, 'name': CloneVM_Task, 'duration_secs': 1.801587} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.034945] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Created linked-clone VM from snapshot [ 751.037211] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4b74dc-8951-48fe-98ff-d7533e9527ec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.049430] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Uploading image 57e934b3-5f97-43b1-a6e2-ce5757d3ed70 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 751.077987] env[61728]: DEBUG oslo_vmware.rw_handles [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 751.077987] env[61728]: value = "vm-122081" [ 751.077987] env[61728]: _type = "VirtualMachine" [ 751.077987] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 751.078284] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b1e9f99e-00ad-4284-8bf2-a27f9e10b4d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.109142] env[61728]: DEBUG oslo_vmware.rw_handles [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lease: (returnval){ [ 751.109142] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52233ddd-c8db-330e-8996-a01f814c61c0" [ 751.109142] env[61728]: _type = "HttpNfcLease" [ 751.109142] env[61728]: } obtained for exporting VM: (result){ [ 751.109142] env[61728]: value = "vm-122081" [ 751.109142] env[61728]: _type = "VirtualMachine" [ 751.109142] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 751.110097] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the lease: (returnval){ [ 751.110097] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52233ddd-c8db-330e-8996-a01f814c61c0" [ 751.110097] env[61728]: _type = "HttpNfcLease" [ 751.110097] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 751.121488] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 751.121488] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52233ddd-c8db-330e-8996-a01f814c61c0" [ 751.121488] env[61728]: _type = "HttpNfcLease" [ 751.121488] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 751.123453] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22fb1674-8d59-4eea-ae38-f9a27099e8d2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.132843] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf89122f-7474-4e9c-bdcb-2c8d02909448 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.171162] env[61728]: DEBUG nova.network.neutron [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Updating instance_info_cache with network_info: [{"id": "3f69dcc2-f3d8-4b50-b760-85f6849a6c24", "address": "fa:16:3e:77:1c:7f", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f69dcc2-f3", "ovs_interfaceid": "3f69dcc2-f3d8-4b50-b760-85f6849a6c24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.175193] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd2fced-3635-4412-b0b6-c264fca61495 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.185542] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3167b128-bd70-4e20-afe7-148d9512e3d8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.202012] env[61728]: DEBUG nova.compute.provider_tree [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.222753] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.222854] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.223558] env[61728]: DEBUG nova.network.neutron [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 751.261778] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464071, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.341449] env[61728]: DEBUG nova.network.neutron [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance_info_cache with network_info: [{"id": "6c9a6b61-5248-4996-a390-c18edb3ec021", "address": "fa:16:3e:43:9a:d5", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9a6b61-52", "ovs_interfaceid": "6c9a6b61-5248-4996-a390-c18edb3ec021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.601060] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "b563e521-0e0e-45ea-bd5a-056285ea35b9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.601317] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "b563e521-0e0e-45ea-bd5a-056285ea35b9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.601643] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "b563e521-0e0e-45ea-bd5a-056285ea35b9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.601908] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "b563e521-0e0e-45ea-bd5a-056285ea35b9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.602167] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "b563e521-0e0e-45ea-bd5a-056285ea35b9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.604976] env[61728]: INFO nova.compute.manager [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Terminating instance [ 751.607673] env[61728]: DEBUG nova.compute.manager [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 751.607957] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 751.609129] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5394de0-b8e0-450c-8555-871cd16bc063 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.622024] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 751.622024] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52233ddd-c8db-330e-8996-a01f814c61c0" [ 751.622024] env[61728]: _type = "HttpNfcLease" [ 751.622024] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 751.624896] env[61728]: DEBUG oslo_vmware.rw_handles [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 751.624896] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52233ddd-c8db-330e-8996-a01f814c61c0" [ 751.624896] env[61728]: _type = "HttpNfcLease" [ 751.624896] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 751.625294] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 751.626401] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0e2856-f63c-42e2-88ff-21f3a7064d0f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.629525] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c488e2b9-3b01-41cc-bd72-300fdbc3bc5e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.638521] env[61728]: DEBUG oslo_vmware.rw_handles [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524c6094-0124-ecd0-3420-37dac255176c/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 751.638813] env[61728]: DEBUG oslo_vmware.rw_handles [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524c6094-0124-ecd0-3420-37dac255176c/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 751.641793] env[61728]: DEBUG oslo_vmware.api [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 751.641793] env[61728]: value = "task-464073" [ 751.641793] env[61728]: _type = "Task" [ 751.641793] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.724224] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "refresh_cache-40ef29df-12d6-4274-846a-4a2c32752b8b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.724680] env[61728]: DEBUG nova.compute.manager [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Instance network_info: |[{"id": "3f69dcc2-f3d8-4b50-b760-85f6849a6c24", "address": "fa:16:3e:77:1c:7f", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f69dcc2-f3", "ovs_interfaceid": "3f69dcc2-f3d8-4b50-b760-85f6849a6c24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 751.725835] env[61728]: DEBUG nova.scheduler.client.report [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.732604] env[61728]: DEBUG oslo_concurrency.lockutils [req-e4e3d336-6ec3-40b6-98b5-35aae39ea9a7 req-203434d6-cca7-4558-bf5d-22e6494d8dc9 service nova] Acquired lock "refresh_cache-40ef29df-12d6-4274-846a-4a2c32752b8b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.732938] env[61728]: DEBUG nova.network.neutron [req-e4e3d336-6ec3-40b6-98b5-35aae39ea9a7 req-203434d6-cca7-4558-bf5d-22e6494d8dc9 service nova] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Refreshing network info cache for port 3f69dcc2-f3d8-4b50-b760-85f6849a6c24 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 751.737316] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:1c:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc6d5964-1106-4345-a26d-185dabd4ff0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3f69dcc2-f3d8-4b50-b760-85f6849a6c24', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 751.746801] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Creating folder: Project (3a02bd852c564acd99e12fd17279f028). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 751.755807] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7c7668d2-bf63-4667-a6b1-fa02072965c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.757783] env[61728]: DEBUG oslo_vmware.api [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-464073, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.773390] env[61728]: DEBUG oslo_vmware.api [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464071, 'name': PowerOnVM_Task, 'duration_secs': 0.763488} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.774999] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 751.775302] env[61728]: INFO nova.compute.manager [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Took 6.39 seconds to spawn the instance on the hypervisor. [ 751.776307] env[61728]: DEBUG nova.compute.manager [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 751.776307] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Created folder: Project (3a02bd852c564acd99e12fd17279f028) in parent group-v121913. [ 751.776307] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Creating folder: Instances. Parent ref: group-v122082. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 751.776752] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bbc255e-69ac-4a21-b2f0-f128b29b821a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.779624] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-06ec48a5-5ebe-48f0-8be5-2afaf2d01149 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.792327] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5a9b706e-3501-4e0d-b46b-3d9278df5e28 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.797541] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Created folder: Instances in parent group-v122082. [ 751.797808] env[61728]: DEBUG oslo.service.loopingcall [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.799415] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 751.800754] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7aed6c31-938a-4a6d-b115-e81a6db0a7b5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.824789] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 751.824789] env[61728]: value = "task-464076" [ 751.824789] env[61728]: _type = "Task" [ 751.824789] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.832051] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464076, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.847644] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.106232] env[61728]: DEBUG nova.network.neutron [req-e4e3d336-6ec3-40b6-98b5-35aae39ea9a7 req-203434d6-cca7-4558-bf5d-22e6494d8dc9 service nova] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Updated VIF entry in instance network info cache for port 3f69dcc2-f3d8-4b50-b760-85f6849a6c24. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 752.106688] env[61728]: DEBUG nova.network.neutron [req-e4e3d336-6ec3-40b6-98b5-35aae39ea9a7 req-203434d6-cca7-4558-bf5d-22e6494d8dc9 service nova] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Updating instance_info_cache with network_info: [{"id": "3f69dcc2-f3d8-4b50-b760-85f6849a6c24", "address": "fa:16:3e:77:1c:7f", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f69dcc2-f3", "ovs_interfaceid": "3f69dcc2-f3d8-4b50-b760-85f6849a6c24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.149426] env[61728]: DEBUG nova.network.neutron [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance_info_cache with network_info: [{"id": "4d70757b-d260-4aa3-8914-a38eb567f7b8", "address": "fa:16:3e:2a:f5:67", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.112", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d70757b-d2", "ovs_interfaceid": "4d70757b-d260-4aa3-8914-a38eb567f7b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.156897] env[61728]: DEBUG oslo_vmware.api [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-464073, 'name': PowerOffVM_Task, 'duration_secs': 0.323437} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.157234] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 752.157423] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 752.157720] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cde04d4e-1167-423b-9556-8a0386e22a0a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.222967] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 752.223959] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 752.223959] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Deleting the datastore file [datastore1] b563e521-0e0e-45ea-bd5a-056285ea35b9 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 752.224514] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc17a8e9-4cb4-427a-8f33-df45faa92a3a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.232445] env[61728]: DEBUG oslo_vmware.api [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for the task: (returnval){ [ 752.232445] env[61728]: value = "task-464078" [ 752.232445] env[61728]: _type = "Task" [ 752.232445] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.242598] env[61728]: DEBUG oslo_vmware.api [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-464078, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.254889] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.210s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.257841] env[61728]: DEBUG oslo_concurrency.lockutils [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.508s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.258333] env[61728]: DEBUG nova.objects.instance [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Lazy-loading 'resources' on Instance uuid 8c53c9c9-cdc1-407b-8827-8409cf137235 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 752.279815] env[61728]: INFO nova.scheduler.client.report [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Deleted allocations for instance ca8d41dc-5719-4ade-b82a-b1d7f1c8def1 [ 752.303641] env[61728]: INFO nova.compute.manager [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Took 51.82 seconds to build instance. [ 752.337027] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464076, 'name': CreateVM_Task, 'duration_secs': 0.360457} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.337027] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 752.337027] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.337027] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.337027] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 752.337027] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b53c31f4-b059-4081-a103-2d797e117988 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.344143] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 752.344143] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524fa135-aba0-75f3-76f2-339a48afabbf" [ 752.344143] env[61728]: _type = "Task" [ 752.344143] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.362071] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524fa135-aba0-75f3-76f2-339a48afabbf, 'name': SearchDatastore_Task, 'duration_secs': 0.010412} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.363301] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.363301] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 752.363764] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.364100] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.364453] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 752.364847] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4689d42a-4892-42ef-b262-f51a44472325 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.375567] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 752.376442] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 752.377714] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ad29f6c-3bce-4d66-a9b8-798f27a24a32 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.386478] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 752.386478] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527873ed-fbd5-9d60-7725-b9d5bc8dadcd" [ 752.386478] env[61728]: _type = "Task" [ 752.386478] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.400639] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527873ed-fbd5-9d60-7725-b9d5bc8dadcd, 'name': SearchDatastore_Task, 'duration_secs': 0.011122} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.401632] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb40c6fa-de04-41f6-9253-247451909344 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.409443] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 752.409443] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52108b5e-d12a-3717-1115-19cd15655f34" [ 752.409443] env[61728]: _type = "Task" [ 752.409443] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.418196] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52108b5e-d12a-3717-1115-19cd15655f34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.613449] env[61728]: DEBUG oslo_concurrency.lockutils [req-e4e3d336-6ec3-40b6-98b5-35aae39ea9a7 req-203434d6-cca7-4558-bf5d-22e6494d8dc9 service nova] Releasing lock "refresh_cache-40ef29df-12d6-4274-846a-4a2c32752b8b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.652387] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Releasing lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.744873] env[61728]: DEBUG oslo_vmware.api [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Task: {'id': task-464078, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.229598} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.745160] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 752.745355] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 752.745536] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 752.745723] env[61728]: INFO nova.compute.manager [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Took 1.14 seconds to destroy the instance on the hypervisor. [ 752.746027] env[61728]: DEBUG oslo.service.loopingcall [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.746264] env[61728]: DEBUG nova.compute.manager [-] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.746370] env[61728]: DEBUG nova.network.neutron [-] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 752.789305] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5744e707-5ca1-4d80-89e8-368a33470b40 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "ca8d41dc-5719-4ade-b82a-b1d7f1c8def1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 45.865s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.805679] env[61728]: DEBUG oslo_concurrency.lockutils [None req-849556fd-1adf-4b57-981d-cb0617b95fbd tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "2e83145a-2481-411d-b037-a305a209074e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.898s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.922932] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52108b5e-d12a-3717-1115-19cd15655f34, 'name': SearchDatastore_Task, 'duration_secs': 0.010485} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.926198] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.926495] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 40ef29df-12d6-4274-846a-4a2c32752b8b/40ef29df-12d6-4274-846a-4a2c32752b8b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 752.927032] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb817af2-83a6-4ca2-a2fd-e05df861a7b4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.939127] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 752.939127] env[61728]: value = "task-464079" [ 752.939127] env[61728]: _type = "Task" [ 752.939127] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.950339] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464079, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.185382] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6661fd6-2fa3-42da-beb0-3647029635a8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.216118] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54f63000-32a7-494e-b876-d9039e28c741 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.225770] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance '1d9114cb-a0af-4874-962f-27237b3c89cc' progress to 83 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 753.313104] env[61728]: DEBUG nova.compute.manager [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 753.365640] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14eb47d5-5dbd-43e8-9070-f96aa52ebc11 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.397383] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance '7b16fbc0-7f13-405f-b84e-e18de1ca7dd2' progress to 0 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 753.452908] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464079, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.512078} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.454270] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 40ef29df-12d6-4274-846a-4a2c32752b8b/40ef29df-12d6-4274-846a-4a2c32752b8b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 753.454629] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 753.455035] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-40857a5f-c238-43d0-a3c2-c5f159908bb3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.467444] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 753.467444] env[61728]: value = "task-464080" [ 753.467444] env[61728]: _type = "Task" [ 753.467444] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.479705] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464080, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.594743] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "3ef9bf60-c6a1-4b7a-a375-5397fb871850" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.595451] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "3ef9bf60-c6a1-4b7a-a375-5397fb871850" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.609593] env[61728]: DEBUG nova.compute.manager [req-fd83dc60-abbd-49d0-8cf3-d3d5cf9a6fd6 req-db1f3894-1289-4322-8b9b-5f8c44d0bf91 service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Received event network-vif-deleted-613976dd-b5d2-4919-9887-b3d42dc84c87 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 753.609846] env[61728]: INFO nova.compute.manager [req-fd83dc60-abbd-49d0-8cf3-d3d5cf9a6fd6 req-db1f3894-1289-4322-8b9b-5f8c44d0bf91 service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Neutron deleted interface 613976dd-b5d2-4919-9887-b3d42dc84c87; detaching it from the instance and deleting it from the info cache [ 753.610475] env[61728]: DEBUG nova.network.neutron [req-fd83dc60-abbd-49d0-8cf3-d3d5cf9a6fd6 req-db1f3894-1289-4322-8b9b-5f8c44d0bf91 service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.613320] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49538c60-442d-4bb4-9c36-1b028e26634b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.623113] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf373cd-a798-4e5c-8b0e-dde4d53b1898 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.659855] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3cca717-a783-4f4a-94ef-e340743130d9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.672213] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06dbe06b-9c5c-4eb6-99da-3721ffc39b75 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.690451] env[61728]: DEBUG nova.compute.provider_tree [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.733821] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 753.734447] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec4bf70e-d581-4e1c-9598-ff9ee2d2bff0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.743292] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 753.743292] env[61728]: value = "task-464081" [ 753.743292] env[61728]: _type = "Task" [ 753.743292] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.753656] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464081, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.783513] env[61728]: DEBUG nova.network.neutron [-] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.832402] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.844447] env[61728]: INFO nova.compute.manager [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Rebuilding instance [ 753.899743] env[61728]: DEBUG nova.compute.manager [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 753.900679] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b91195b-a3c3-41e2-b420-629683031f9a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.905330] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 753.905591] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80dcaed1-67c8-4d1c-b93b-0da1c6863523 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.916408] env[61728]: DEBUG oslo_vmware.api [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 753.916408] env[61728]: value = "task-464082" [ 753.916408] env[61728]: _type = "Task" [ 753.916408] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.926671] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 753.926935] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance '7b16fbc0-7f13-405f-b84e-e18de1ca7dd2' progress to 17 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 753.983265] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464080, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.157022} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.984293] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 753.985117] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe76eff-6879-461b-a4f5-582a7ed6fb7c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.010636] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 40ef29df-12d6-4274-846a-4a2c32752b8b/40ef29df-12d6-4274-846a-4a2c32752b8b.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 754.011336] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01097262-5237-4061-9d8f-818d266d0895 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.033492] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 754.033492] env[61728]: value = "task-464083" [ 754.033492] env[61728]: _type = "Task" [ 754.033492] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.043575] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464083, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.118354] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5a99a44a-e61e-458e-aca9-e4212b2e581b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.130203] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b14b285-fd48-4f7d-834c-ca48af0fe832 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.176381] env[61728]: DEBUG nova.compute.manager [req-fd83dc60-abbd-49d0-8cf3-d3d5cf9a6fd6 req-db1f3894-1289-4322-8b9b-5f8c44d0bf91 service nova] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Detach interface failed, port_id=613976dd-b5d2-4919-9887-b3d42dc84c87, reason: Instance b563e521-0e0e-45ea-bd5a-056285ea35b9 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 754.193971] env[61728]: DEBUG nova.scheduler.client.report [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.255406] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464081, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.287012] env[61728]: INFO nova.compute.manager [-] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Took 1.54 seconds to deallocate network for instance. [ 754.418468] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 754.418895] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dfa3295f-b85a-486f-ae61-e341fadb8d55 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.427743] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 754.427743] env[61728]: value = "task-464084" [ 754.427743] env[61728]: _type = "Task" [ 754.427743] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.434521] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 754.434952] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 754.435251] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 754.435572] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 754.435845] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 754.436140] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 754.436505] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 754.436737] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 754.437046] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 754.437355] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 754.437680] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 754.446329] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed169dd3-5faf-485f-b009-13d74da9b4f3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.470806] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464084, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.472855] env[61728]: DEBUG oslo_vmware.api [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 754.472855] env[61728]: value = "task-464085" [ 754.472855] env[61728]: _type = "Task" [ 754.472855] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.485021] env[61728]: DEBUG oslo_vmware.api [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464085, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.544643] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464083, 'name': ReconfigVM_Task, 'duration_secs': 0.300568} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.544956] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 40ef29df-12d6-4274-846a-4a2c32752b8b/40ef29df-12d6-4274-846a-4a2c32752b8b.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 754.545618] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0ac5a003-3f45-4879-ae81-dbf2a7bce4a7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.553057] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 754.553057] env[61728]: value = "task-464086" [ 754.553057] env[61728]: _type = "Task" [ 754.553057] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.563525] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464086, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.699375] env[61728]: DEBUG oslo_concurrency.lockutils [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.441s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.702104] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 41.262s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.702318] env[61728]: DEBUG nova.objects.instance [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61728) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 754.720155] env[61728]: INFO nova.scheduler.client.report [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Deleted allocations for instance 8c53c9c9-cdc1-407b-8827-8409cf137235 [ 754.756854] env[61728]: DEBUG oslo_vmware.api [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464081, 'name': PowerOnVM_Task, 'duration_secs': 0.571638} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.757326] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 754.757550] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b0051e5e-c3c5-4f04-b7fc-f81cc80d2490 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance '1d9114cb-a0af-4874-962f-27237b3c89cc' progress to 100 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 754.794900] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.938436] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464084, 'name': PowerOffVM_Task, 'duration_secs': 0.217253} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.938788] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 754.938946] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 754.939766] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db167483-26d2-4ad3-a3b1-94ef0d1f1157 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.947325] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 754.947564] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8c0e8138-91ea-467f-9a33-1b57e6f1e6b2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.979791] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 754.979842] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 754.980055] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Deleting the datastore file [datastore1] 2e83145a-2481-411d-b037-a305a209074e {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 754.980829] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a50cf095-7633-4c6c-b476-4fc8fb810d9b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.988500] env[61728]: DEBUG oslo_vmware.api [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464085, 'name': ReconfigVM_Task, 'duration_secs': 0.177041} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.989879] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance '7b16fbc0-7f13-405f-b84e-e18de1ca7dd2' progress to 33 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 754.993781] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 754.993781] env[61728]: value = "task-464088" [ 754.993781] env[61728]: _type = "Task" [ 754.993781] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.005350] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464088, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.064714] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464086, 'name': Rename_Task, 'duration_secs': 0.16791} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.065037] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 755.065345] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f68ef00-8178-441d-9819-fd1707e5b054 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.074263] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 755.074263] env[61728]: value = "task-464089" [ 755.074263] env[61728]: _type = "Task" [ 755.074263] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.084055] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464089, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.232412] env[61728]: DEBUG oslo_concurrency.lockutils [None req-83793970-a69b-4b72-8ab8-581de2570491 tempest-ServersTestFqdnHostnames-665258478 tempest-ServersTestFqdnHostnames-665258478-project-member] Lock "8c53c9c9-cdc1-407b-8827-8409cf137235" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 45.985s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.497051] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 755.497347] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 755.497515] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 755.497811] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 755.497898] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 755.498011] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 755.498291] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 755.498389] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 755.498563] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 755.498733] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 755.498916] env[61728]: DEBUG nova.virt.hardware [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.505323] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Reconfiguring VM instance instance-00000018 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 755.505756] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-738d8735-ffac-4d18-96ea-e68054aef6ce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.536022] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464088, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10074} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.537512] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 755.537742] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 755.537930] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 755.540495] env[61728]: DEBUG oslo_vmware.api [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 755.540495] env[61728]: value = "task-464090" [ 755.540495] env[61728]: _type = "Task" [ 755.540495] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.553403] env[61728]: DEBUG oslo_vmware.api [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464090, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.587339] env[61728]: DEBUG oslo_vmware.api [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464089, 'name': PowerOnVM_Task, 'duration_secs': 0.499177} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.587339] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 755.587339] env[61728]: INFO nova.compute.manager [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Took 6.31 seconds to spawn the instance on the hypervisor. [ 755.587339] env[61728]: DEBUG nova.compute.manager [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 755.587339] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b6aa94-309e-4676-be65-037d79489980 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.716607] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ba39d82d-7a25-4fa3-91dc-2a5855335d7b tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.717761] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.819s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.720167] env[61728]: INFO nova.compute.claims [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.053924] env[61728]: DEBUG oslo_vmware.api [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464090, 'name': ReconfigVM_Task, 'duration_secs': 0.301956} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.055430] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Reconfigured VM instance instance-00000018 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 756.055430] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd42306c-6f0c-4fcf-b3be-85a06374f514 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.082224] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Reconfiguring VM instance instance-00000018 to attach disk [datastore1] 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2/7b16fbc0-7f13-405f-b84e-e18de1ca7dd2.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 756.083025] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b008a2c3-f8b6-4601-b36e-d2b618fa95fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.107864] env[61728]: DEBUG oslo_vmware.api [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 756.107864] env[61728]: value = "task-464091" [ 756.107864] env[61728]: _type = "Task" [ 756.107864] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.113169] env[61728]: INFO nova.compute.manager [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Took 53.70 seconds to build instance. [ 756.122188] env[61728]: DEBUG oslo_vmware.api [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464091, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.586127] env[61728]: DEBUG nova.virt.hardware [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 756.586127] env[61728]: DEBUG nova.virt.hardware [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 756.586127] env[61728]: DEBUG nova.virt.hardware [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 756.586127] env[61728]: DEBUG nova.virt.hardware [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 756.586936] env[61728]: DEBUG nova.virt.hardware [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 756.588011] env[61728]: DEBUG nova.virt.hardware [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 756.588011] env[61728]: DEBUG nova.virt.hardware [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 756.588011] env[61728]: DEBUG nova.virt.hardware [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 756.588011] env[61728]: DEBUG nova.virt.hardware [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 756.588011] env[61728]: DEBUG nova.virt.hardware [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 756.588863] env[61728]: DEBUG nova.virt.hardware [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 756.589854] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ea25d6-7bee-402c-8e1e-52135bb0ec4b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.601521] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8bd6dc0-cb39-4355-ae50-bd6371c8ba72 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.624534] env[61728]: DEBUG oslo_concurrency.lockutils [None req-faf028b4-2168-4f00-b344-0a8ae89ee06c tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "40ef29df-12d6-4274-846a-4a2c32752b8b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.657s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.625015] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Instance VIF info [] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 756.630802] env[61728]: DEBUG oslo.service.loopingcall [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 756.635389] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e83145a-2481-411d-b037-a305a209074e] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 756.635852] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c6e1187f-d1e6-429b-bef9-8ef711c34139 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.655052] env[61728]: DEBUG oslo_vmware.api [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464091, 'name': ReconfigVM_Task, 'duration_secs': 0.348318} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.656482] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Reconfigured VM instance instance-00000018 to attach disk [datastore1] 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2/7b16fbc0-7f13-405f-b84e-e18de1ca7dd2.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 756.656781] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance '7b16fbc0-7f13-405f-b84e-e18de1ca7dd2' progress to 50 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 756.660758] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 756.660758] env[61728]: value = "task-464092" [ 756.660758] env[61728]: _type = "Task" [ 756.660758] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.680030] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464092, 'name': CreateVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.136988] env[61728]: DEBUG nova.compute.manager [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.170487] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6defb7b-9329-421b-a282-6142ff6aa693 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.189534] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464092, 'name': CreateVM_Task, 'duration_secs': 0.4487} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.191664] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e83145a-2481-411d-b037-a305a209074e] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 757.205885] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.206078] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.206410] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 757.210120] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a04aa273-3766-45f4-aca2-b3e1e43c4687 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.212385] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3adda17-cfbd-45dc-8192-721ca0b3f861 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.234109] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance '7b16fbc0-7f13-405f-b84e-e18de1ca7dd2' progress to 67 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 757.238173] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 757.238173] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b1ef04-a1f5-259f-77ca-371742f00903" [ 757.238173] env[61728]: _type = "Task" [ 757.238173] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.254326] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b1ef04-a1f5-259f-77ca-371742f00903, 'name': SearchDatastore_Task, 'duration_secs': 0.028025} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.255039] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.255039] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 757.255183] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.255556] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.255556] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 757.258093] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca11e7d9-f8db-4f69-a839-b99db210cd6e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.268712] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 757.270497] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 757.270497] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d3fc231-665c-4167-b02b-51948a67fdaa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.278068] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 757.278068] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f9494f-2898-ef50-16d5-82074e657527" [ 757.278068] env[61728]: _type = "Task" [ 757.278068] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.291449] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f9494f-2898-ef50-16d5-82074e657527, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.353375] env[61728]: DEBUG nova.network.neutron [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Port 4d70757b-d260-4aa3-8914-a38eb567f7b8 binding to destination host cpu-1 is already ACTIVE {{(pid=61728) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 757.354027] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.354027] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.354027] env[61728]: DEBUG nova.network.neutron [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 757.474962] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a65d950f-61ce-41dd-8ffc-10518cc9b953 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.485027] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2791014a-46bb-4670-969c-cc6d3c863fcd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.519823] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d653d45d-2d00-4563-b74f-6a4544dc7e97 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.528279] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0458755-14f2-4515-b1a7-f5e0c2d0804f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.543611] env[61728]: DEBUG nova.compute.provider_tree [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.659234] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.779272] env[61728]: DEBUG nova.network.neutron [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Port 6c9a6b61-5248-4996-a390-c18edb3ec021 binding to destination host cpu-1 is already ACTIVE {{(pid=61728) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 757.791799] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f9494f-2898-ef50-16d5-82074e657527, 'name': SearchDatastore_Task, 'duration_secs': 0.020121} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.793608] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8b0f60f-5257-484a-af66-0e2d3b8b57ee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.800500] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 757.800500] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5262cda6-ce63-c59f-f046-7ab7af3cfa94" [ 757.800500] env[61728]: _type = "Task" [ 757.800500] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.810180] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5262cda6-ce63-c59f-f046-7ab7af3cfa94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.048844] env[61728]: DEBUG nova.scheduler.client.report [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.104813] env[61728]: DEBUG nova.network.neutron [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance_info_cache with network_info: [{"id": "4d70757b-d260-4aa3-8914-a38eb567f7b8", "address": "fa:16:3e:2a:f5:67", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.112", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d70757b-d2", "ovs_interfaceid": "4d70757b-d260-4aa3-8914-a38eb567f7b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.318827] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5262cda6-ce63-c59f-f046-7ab7af3cfa94, 'name': SearchDatastore_Task, 'duration_secs': 0.04299} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.319650] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.319650] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2e83145a-2481-411d-b037-a305a209074e/2e83145a-2481-411d-b037-a305a209074e.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 758.319967] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7101a163-e4bf-40ee-9376-34d8e6be4699 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.329155] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 758.329155] env[61728]: value = "task-464093" [ 758.329155] env[61728]: _type = "Task" [ 758.329155] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.339465] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464093, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.557405] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.839s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.558721] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.785s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.561386] env[61728]: INFO nova.compute.claims [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 758.608617] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Releasing lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.674843] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "d44718bd-872b-401c-aa11-f10bea4a35d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.674843] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "d44718bd-872b-401c-aa11-f10bea4a35d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.805431] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.805640] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.805880] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.841957] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464093, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.068916] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Acquiring lock "da46a4c7-11bc-4ad8-83f0-bad594c6a0a4" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.069217] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Lock "da46a4c7-11bc-4ad8-83f0-bad594c6a0a4" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.114687] env[61728]: DEBUG nova.compute.manager [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61728) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 759.114868] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.340692] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464093, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.627815} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.340984] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2e83145a-2481-411d-b037-a305a209074e/2e83145a-2481-411d-b037-a305a209074e.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 759.341231] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 759.341488] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-76393435-1df1-4132-a50e-d8802b8b5b02 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.350559] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 759.350559] env[61728]: value = "task-464094" [ 759.350559] env[61728]: _type = "Task" [ 759.350559] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.361826] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464094, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.574138] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Lock "da46a4c7-11bc-4ad8-83f0-bad594c6a0a4" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.574740] env[61728]: DEBUG nova.compute.manager [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 759.847552] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.848116] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.848116] env[61728]: DEBUG nova.network.neutron [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 759.862865] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464094, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.222901} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.865758] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 759.867476] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1c9aa7-424f-48b1-94b9-5091d558200b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.888750] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 2e83145a-2481-411d-b037-a305a209074e/2e83145a-2481-411d-b037-a305a209074e.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 759.891707] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fde04bc5-20b2-4feb-a992-9a4037284053 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.912982] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 759.912982] env[61728]: value = "task-464095" [ 759.912982] env[61728]: _type = "Task" [ 759.912982] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.924247] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464095, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.081314] env[61728]: DEBUG nova.compute.utils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.083067] env[61728]: DEBUG nova.compute.manager [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 760.083246] env[61728]: DEBUG nova.network.neutron [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 760.108804] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c57d76b-f366-4eb3-85b9-b1d55df56aed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.117945] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93cd84a1-591e-41c0-92eb-e45276788131 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.151372] env[61728]: DEBUG nova.policy [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3d45dd014b44e02926d6b8282f99950', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e40475ffd52a4413b0f7db4a6b42c68b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 760.153548] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28577ef4-d169-46c0-86f7-e92304e3ea82 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.162232] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02fc01a8-5219-4b9e-9361-b1be62b6512e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.177274] env[61728]: DEBUG nova.compute.provider_tree [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.430347] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464095, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.458142] env[61728]: DEBUG nova.network.neutron [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Successfully created port: 9666e319-339d-4b02-93fa-ec1aaf79fb7b {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 760.586179] env[61728]: DEBUG nova.compute.manager [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 760.680714] env[61728]: DEBUG nova.scheduler.client.report [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.928309] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464095, 'name': ReconfigVM_Task, 'duration_secs': 0.772719} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.928665] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 2e83145a-2481-411d-b037-a305a209074e/2e83145a-2481-411d-b037-a305a209074e.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 760.929307] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b81f002c-e519-43fa-978e-7cbd52b42213 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.935028] env[61728]: DEBUG nova.network.neutron [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance_info_cache with network_info: [{"id": "6c9a6b61-5248-4996-a390-c18edb3ec021", "address": "fa:16:3e:43:9a:d5", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9a6b61-52", "ovs_interfaceid": "6c9a6b61-5248-4996-a390-c18edb3ec021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.937981] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 760.937981] env[61728]: value = "task-464096" [ 760.937981] env[61728]: _type = "Task" [ 760.937981] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.948753] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464096, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.187201] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.628s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.187830] env[61728]: DEBUG nova.compute.manager [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 761.192203] env[61728]: DEBUG oslo_concurrency.lockutils [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 42.266s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.192510] env[61728]: DEBUG nova.objects.instance [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lazy-loading 'resources' on Instance uuid b833f691-e76e-4a2e-94a4-7594fadc3ba8 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 761.444026] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.454476] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464096, 'name': Rename_Task, 'duration_secs': 0.187421} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.454846] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 761.455119] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-badfacaa-7f97-46fc-892c-437f4586ff42 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.463494] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 761.463494] env[61728]: value = "task-464097" [ 761.463494] env[61728]: _type = "Task" [ 761.463494] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.472817] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464097, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.595732] env[61728]: DEBUG nova.compute.manager [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 761.620616] env[61728]: DEBUG nova.virt.hardware [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 761.621044] env[61728]: DEBUG nova.virt.hardware [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 761.621253] env[61728]: DEBUG nova.virt.hardware [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 761.621534] env[61728]: DEBUG nova.virt.hardware [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 761.621915] env[61728]: DEBUG nova.virt.hardware [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 761.622156] env[61728]: DEBUG nova.virt.hardware [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 761.622401] env[61728]: DEBUG nova.virt.hardware [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 761.622657] env[61728]: DEBUG nova.virt.hardware [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 761.622906] env[61728]: DEBUG nova.virt.hardware [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 761.623137] env[61728]: DEBUG nova.virt.hardware [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 761.623373] env[61728]: DEBUG nova.virt.hardware [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.624464] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97825e9a-da06-4a0c-93f8-4f20079d8f9c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.633722] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc4770b-5853-4e52-b5ff-a579219e860b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.694182] env[61728]: DEBUG nova.compute.utils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 761.695807] env[61728]: DEBUG nova.compute.manager [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 761.696018] env[61728]: DEBUG nova.network.neutron [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 761.741479] env[61728]: DEBUG nova.policy [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd35296efd6b947fba051391fd697764f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '01839050e56e4dbfa6e22471a3c60998', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 761.972907] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fbf8dfe-0d64-40a4-afd8-f07e32f447eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.983075] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464097, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.002282] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524ae1f8-dfb7-4d68-b12b-98c36b5f7973 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.010801] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance '7b16fbc0-7f13-405f-b84e-e18de1ca7dd2' progress to 83 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 762.051877] env[61728]: DEBUG nova.network.neutron [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Successfully created port: 3f4753c4-b820-4467-8f8a-6084733c6e28 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 762.202475] env[61728]: DEBUG nova.compute.manager [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 762.366414] env[61728]: DEBUG nova.compute.manager [req-a6e7e149-2523-4382-903e-5a714dbb42e0 req-f324d60d-1199-4262-8dea-6e494c1b70e2 service nova] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Received event network-vif-plugged-9666e319-339d-4b02-93fa-ec1aaf79fb7b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 762.366648] env[61728]: DEBUG oslo_concurrency.lockutils [req-a6e7e149-2523-4382-903e-5a714dbb42e0 req-f324d60d-1199-4262-8dea-6e494c1b70e2 service nova] Acquiring lock "39eaa6f4-df3f-4928-a1f1-8c861a392e29-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.366853] env[61728]: DEBUG oslo_concurrency.lockutils [req-a6e7e149-2523-4382-903e-5a714dbb42e0 req-f324d60d-1199-4262-8dea-6e494c1b70e2 service nova] Lock "39eaa6f4-df3f-4928-a1f1-8c861a392e29-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.367032] env[61728]: DEBUG oslo_concurrency.lockutils [req-a6e7e149-2523-4382-903e-5a714dbb42e0 req-f324d60d-1199-4262-8dea-6e494c1b70e2 service nova] Lock "39eaa6f4-df3f-4928-a1f1-8c861a392e29-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.367205] env[61728]: DEBUG nova.compute.manager [req-a6e7e149-2523-4382-903e-5a714dbb42e0 req-f324d60d-1199-4262-8dea-6e494c1b70e2 service nova] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] No waiting events found dispatching network-vif-plugged-9666e319-339d-4b02-93fa-ec1aaf79fb7b {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 762.367369] env[61728]: WARNING nova.compute.manager [req-a6e7e149-2523-4382-903e-5a714dbb42e0 req-f324d60d-1199-4262-8dea-6e494c1b70e2 service nova] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Received unexpected event network-vif-plugged-9666e319-339d-4b02-93fa-ec1aaf79fb7b for instance with vm_state building and task_state spawning. [ 762.374033] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089fc106-a88f-4dad-b765-bede537a7517 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.382925] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbed19c2-425c-4990-82fc-6750a442aa1d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.415623] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a22e26-b577-48f2-a824-70c23b8ff232 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.424462] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aaf26a7-d025-426e-90ce-60d6e8960df9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.439117] env[61728]: DEBUG nova.compute.provider_tree [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.475368] env[61728]: DEBUG oslo_vmware.api [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464097, 'name': PowerOnVM_Task, 'duration_secs': 0.643664} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.475857] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 762.475980] env[61728]: DEBUG nova.compute.manager [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 762.476738] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d560b5c-0fb7-4ec4-bb85-b6c33e716970 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.486534] env[61728]: DEBUG nova.network.neutron [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Successfully updated port: 9666e319-339d-4b02-93fa-ec1aaf79fb7b {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 762.521423] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a81293e8-4999-4291-a668-626aec52652c tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance '7b16fbc0-7f13-405f-b84e-e18de1ca7dd2' progress to 100 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 762.943211] env[61728]: DEBUG nova.scheduler.client.report [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.992340] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Acquiring lock "refresh_cache-39eaa6f4-df3f-4928-a1f1-8c861a392e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.992685] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Acquired lock "refresh_cache-39eaa6f4-df3f-4928-a1f1-8c861a392e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.992685] env[61728]: DEBUG nova.network.neutron [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 762.997520] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.211647] env[61728]: DEBUG nova.compute.manager [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 763.238143] env[61728]: DEBUG nova.virt.hardware [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 763.238455] env[61728]: DEBUG nova.virt.hardware [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 763.238636] env[61728]: DEBUG nova.virt.hardware [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 763.238849] env[61728]: DEBUG nova.virt.hardware [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 763.239015] env[61728]: DEBUG nova.virt.hardware [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 763.239190] env[61728]: DEBUG nova.virt.hardware [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 763.239402] env[61728]: DEBUG nova.virt.hardware [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 763.239643] env[61728]: DEBUG nova.virt.hardware [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 763.239886] env[61728]: DEBUG nova.virt.hardware [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 763.240093] env[61728]: DEBUG nova.virt.hardware [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 763.240297] env[61728]: DEBUG nova.virt.hardware [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 763.241212] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ec9f6a-617c-49a2-8088-365e507960a1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.249983] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd3e5f1-b2f4-4193-b46d-21b13c084e77 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.451023] env[61728]: DEBUG oslo_concurrency.lockutils [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.257s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.452846] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 43.845s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.455617] env[61728]: INFO nova.compute.claims [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 763.484091] env[61728]: INFO nova.scheduler.client.report [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleted allocations for instance b833f691-e76e-4a2e-94a4-7594fadc3ba8 [ 763.553997] env[61728]: DEBUG nova.network.neutron [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 763.654457] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "2e83145a-2481-411d-b037-a305a209074e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.654457] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "2e83145a-2481-411d-b037-a305a209074e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.654668] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "2e83145a-2481-411d-b037-a305a209074e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.654929] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "2e83145a-2481-411d-b037-a305a209074e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.655139] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "2e83145a-2481-411d-b037-a305a209074e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.658322] env[61728]: INFO nova.compute.manager [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Terminating instance [ 763.660405] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "refresh_cache-2e83145a-2481-411d-b037-a305a209074e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.660673] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquired lock "refresh_cache-2e83145a-2481-411d-b037-a305a209074e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.660752] env[61728]: DEBUG nova.network.neutron [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 763.785786] env[61728]: DEBUG nova.network.neutron [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Updating instance_info_cache with network_info: [{"id": "9666e319-339d-4b02-93fa-ec1aaf79fb7b", "address": "fa:16:3e:5f:73:be", "network": {"id": "2a660304-be8d-417c-9db9-aa5f754d7b51", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1542143747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e40475ffd52a4413b0f7db4a6b42c68b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92e4d027-e755-417b-8eea-9a8f24b85140", "external-id": "nsx-vlan-transportzone-756", "segmentation_id": 756, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9666e319-33", "ovs_interfaceid": "9666e319-339d-4b02-93fa-ec1aaf79fb7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.894765] env[61728]: DEBUG nova.network.neutron [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Successfully updated port: 3f4753c4-b820-4467-8f8a-6084733c6e28 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 763.972516] env[61728]: DEBUG oslo_vmware.rw_handles [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524c6094-0124-ecd0-3420-37dac255176c/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 763.973819] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f6246b-6066-4b2b-8c82-c5ab950e694c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.983405] env[61728]: DEBUG oslo_vmware.rw_handles [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524c6094-0124-ecd0-3420-37dac255176c/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 763.983604] env[61728]: ERROR oslo_vmware.rw_handles [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524c6094-0124-ecd0-3420-37dac255176c/disk-0.vmdk due to incomplete transfer. [ 763.983842] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-fb00d4a0-79b3-41ee-afb6-d963423b0db8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.993339] env[61728]: DEBUG oslo_vmware.rw_handles [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524c6094-0124-ecd0-3420-37dac255176c/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 763.993650] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Uploaded image 57e934b3-5f97-43b1-a6e2-ce5757d3ed70 to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 763.996103] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 763.996656] env[61728]: DEBUG oslo_concurrency.lockutils [None req-68d43a60-0f2e-42f1-b6da-b6acaf6c570a tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "b833f691-e76e-4a2e-94a4-7594fadc3ba8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 48.190s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.997824] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f0cc1db0-3590-4c13-ae68-b2501f1e2ef3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.010221] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 764.010221] env[61728]: value = "task-464098" [ 764.010221] env[61728]: _type = "Task" [ 764.010221] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.025803] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464098, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.180985] env[61728]: DEBUG nova.network.neutron [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.237008] env[61728]: DEBUG nova.network.neutron [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.288154] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Releasing lock "refresh_cache-39eaa6f4-df3f-4928-a1f1-8c861a392e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.288498] env[61728]: DEBUG nova.compute.manager [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Instance network_info: |[{"id": "9666e319-339d-4b02-93fa-ec1aaf79fb7b", "address": "fa:16:3e:5f:73:be", "network": {"id": "2a660304-be8d-417c-9db9-aa5f754d7b51", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1542143747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e40475ffd52a4413b0f7db4a6b42c68b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92e4d027-e755-417b-8eea-9a8f24b85140", "external-id": "nsx-vlan-transportzone-756", "segmentation_id": 756, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9666e319-33", "ovs_interfaceid": "9666e319-339d-4b02-93fa-ec1aaf79fb7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 764.288932] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5f:73:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92e4d027-e755-417b-8eea-9a8f24b85140', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9666e319-339d-4b02-93fa-ec1aaf79fb7b', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 764.297165] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Creating folder: Project (e40475ffd52a4413b0f7db4a6b42c68b). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 764.297458] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-19a639d1-ba9a-4dcb-a789-e05fdd3c7366 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.309899] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Created folder: Project (e40475ffd52a4413b0f7db4a6b42c68b) in parent group-v121913. [ 764.310152] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Creating folder: Instances. Parent ref: group-v122086. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 764.310417] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-791b3e3f-e548-419e-8172-7a1b8378e5cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.321547] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Created folder: Instances in parent group-v122086. [ 764.321824] env[61728]: DEBUG oslo.service.loopingcall [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.322038] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 764.322272] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae5bde9e-6711-490e-96cb-77afc7194e44 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.340776] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "2d389d5e-a53f-424e-90af-5ed05e2f93c7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.341086] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "2d389d5e-a53f-424e-90af-5ed05e2f93c7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.341360] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "2d389d5e-a53f-424e-90af-5ed05e2f93c7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.341559] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "2d389d5e-a53f-424e-90af-5ed05e2f93c7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.341762] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "2d389d5e-a53f-424e-90af-5ed05e2f93c7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.344621] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 764.344621] env[61728]: value = "task-464101" [ 764.344621] env[61728]: _type = "Task" [ 764.344621] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.345115] env[61728]: INFO nova.compute.manager [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Terminating instance [ 764.350148] env[61728]: DEBUG nova.compute.manager [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 764.350360] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 764.351103] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370cf2e1-920c-4ec6-9d3f-487ab6b0dc0b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.358461] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 764.361047] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afc1c7b0-432f-41b7-a1fd-a932bfae3bf9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.362664] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464101, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.368625] env[61728]: DEBUG oslo_vmware.api [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 764.368625] env[61728]: value = "task-464102" [ 764.368625] env[61728]: _type = "Task" [ 764.368625] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.378324] env[61728]: DEBUG oslo_vmware.api [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-464102, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.397377] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "refresh_cache-b63fa3df-317a-46d4-b8e9-74e9e287efde" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.397489] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquired lock "refresh_cache-b63fa3df-317a-46d4-b8e9-74e9e287efde" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.397665] env[61728]: DEBUG nova.network.neutron [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 764.460693] env[61728]: DEBUG nova.compute.manager [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Received event network-changed-9666e319-339d-4b02-93fa-ec1aaf79fb7b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 764.461009] env[61728]: DEBUG nova.compute.manager [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Refreshing instance network info cache due to event network-changed-9666e319-339d-4b02-93fa-ec1aaf79fb7b. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 764.461269] env[61728]: DEBUG oslo_concurrency.lockutils [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] Acquiring lock "refresh_cache-39eaa6f4-df3f-4928-a1f1-8c861a392e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.461505] env[61728]: DEBUG oslo_concurrency.lockutils [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] Acquired lock "refresh_cache-39eaa6f4-df3f-4928-a1f1-8c861a392e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.462156] env[61728]: DEBUG nova.network.neutron [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Refreshing network info cache for port 9666e319-339d-4b02-93fa-ec1aaf79fb7b {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 764.525039] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464098, 'name': Destroy_Task, 'duration_secs': 0.341489} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.528076] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Destroyed the VM [ 764.528472] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 764.529136] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a557d25f-5be0-466e-9718-baeea160a474 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.537593] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 764.537593] env[61728]: value = "task-464103" [ 764.537593] env[61728]: _type = "Task" [ 764.537593] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.552726] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464103, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.569034] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.569365] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.569767] env[61728]: DEBUG nova.compute.manager [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Going to confirm migration 3 {{(pid=61728) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 764.740335] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Releasing lock "refresh_cache-2e83145a-2481-411d-b037-a305a209074e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.741368] env[61728]: DEBUG nova.compute.manager [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 764.741822] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 764.743071] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b079dc0e-f2fc-4d46-a9d9-c64a934a25ee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.753578] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 764.756785] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43e86c07-8018-4e6a-a849-485fce578e07 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.766015] env[61728]: DEBUG oslo_vmware.api [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 764.766015] env[61728]: value = "task-464104" [ 764.766015] env[61728]: _type = "Task" [ 764.766015] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.780094] env[61728]: DEBUG oslo_vmware.api [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464104, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.858585] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464101, 'name': CreateVM_Task, 'duration_secs': 0.356207} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.861438] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 764.862488] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.862744] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.863112] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 764.863408] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-725e65f8-edf9-4103-ab4c-699ad6f2718d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.869344] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Waiting for the task: (returnval){ [ 764.869344] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b5da0f-a3b9-f80b-a0e9-70f564dd2f0f" [ 764.869344] env[61728]: _type = "Task" [ 764.869344] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.888542] env[61728]: DEBUG oslo_vmware.api [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-464102, 'name': PowerOffVM_Task, 'duration_secs': 0.216649} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.888867] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b5da0f-a3b9-f80b-a0e9-70f564dd2f0f, 'name': SearchDatastore_Task, 'duration_secs': 0.011008} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.891766] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 764.891983] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 764.892303] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.892543] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 764.892798] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.892958] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.893206] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 764.893660] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-682973a6-250d-4b8f-a766-05cab99e486f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.895322] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-09e0adbc-3f60-4f04-ae28-519e37f2f1b2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.907536] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 764.908359] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 764.908571] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5121155-5e2c-4679-82cb-79747b1b1d43 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.914546] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Waiting for the task: (returnval){ [ 764.914546] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5277ba2c-a278-7036-0275-7dec314aced7" [ 764.914546] env[61728]: _type = "Task" [ 764.914546] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.928950] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5277ba2c-a278-7036-0275-7dec314aced7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.930786] env[61728]: DEBUG nova.network.neutron [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.978039] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 764.978420] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 764.978531] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleting the datastore file [datastore1] 2d389d5e-a53f-424e-90af-5ed05e2f93c7 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 764.978814] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9a4bb976-0983-4194-aa77-bd82e8554f93 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.990679] env[61728]: DEBUG oslo_vmware.api [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 764.990679] env[61728]: value = "task-464106" [ 764.990679] env[61728]: _type = "Task" [ 764.990679] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.004221] env[61728]: DEBUG oslo_vmware.api [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-464106, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.053045] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464103, 'name': RemoveSnapshot_Task} progress is 44%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.083088] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb5409eb-e17b-4561-afd1-e37251381bf1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.091699] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3f63fd-ef1c-498c-a4f8-be37bbfd91ce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.132921] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a4c345-8127-4d84-b99b-5486cecd9c52 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.143645] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80377e2-d537-4ffa-a56e-87253fde580f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.158067] env[61728]: DEBUG nova.compute.provider_tree [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.216867] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.216867] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.216867] env[61728]: DEBUG nova.network.neutron [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 765.216867] env[61728]: DEBUG nova.objects.instance [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'info_cache' on Instance uuid 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 765.226996] env[61728]: DEBUG nova.network.neutron [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Updating instance_info_cache with network_info: [{"id": "3f4753c4-b820-4467-8f8a-6084733c6e28", "address": "fa:16:3e:ca:57:67", "network": {"id": "6a1908cc-f6aa-4dd7-970e-c8a07e1b16ee", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2047209328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "01839050e56e4dbfa6e22471a3c60998", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f4753c4-b8", "ovs_interfaceid": "3f4753c4-b820-4467-8f8a-6084733c6e28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.280578] env[61728]: DEBUG oslo_vmware.api [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464104, 'name': PowerOffVM_Task, 'duration_secs': 0.218373} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.281043] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 765.281366] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 765.281765] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35bc11b5-6e20-4160-a21f-636a88f30966 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.294756] env[61728]: DEBUG nova.network.neutron [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Updated VIF entry in instance network info cache for port 9666e319-339d-4b02-93fa-ec1aaf79fb7b. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 765.295298] env[61728]: DEBUG nova.network.neutron [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Updating instance_info_cache with network_info: [{"id": "9666e319-339d-4b02-93fa-ec1aaf79fb7b", "address": "fa:16:3e:5f:73:be", "network": {"id": "2a660304-be8d-417c-9db9-aa5f754d7b51", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1542143747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e40475ffd52a4413b0f7db4a6b42c68b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92e4d027-e755-417b-8eea-9a8f24b85140", "external-id": "nsx-vlan-transportzone-756", "segmentation_id": 756, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9666e319-33", "ovs_interfaceid": "9666e319-339d-4b02-93fa-ec1aaf79fb7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.315267] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 765.315709] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 765.315930] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Deleting the datastore file [datastore1] 2e83145a-2481-411d-b037-a305a209074e {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 765.316249] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-48c288df-ab8f-4107-9f87-a61c20ee6fdc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.324099] env[61728]: DEBUG oslo_vmware.api [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 765.324099] env[61728]: value = "task-464108" [ 765.324099] env[61728]: _type = "Task" [ 765.324099] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.333617] env[61728]: DEBUG oslo_vmware.api [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464108, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.426256] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5277ba2c-a278-7036-0275-7dec314aced7, 'name': SearchDatastore_Task, 'duration_secs': 0.01137} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.427255] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8517c1fa-5efc-4121-825e-d16a25d45c11 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.433042] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Waiting for the task: (returnval){ [ 765.433042] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5252841a-5840-4037-3454-d544fc47b2b1" [ 765.433042] env[61728]: _type = "Task" [ 765.433042] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.443878] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5252841a-5840-4037-3454-d544fc47b2b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.502257] env[61728]: DEBUG oslo_vmware.api [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-464106, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148199} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.503349] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 765.503716] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 765.504046] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 765.504452] env[61728]: INFO nova.compute.manager [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Took 1.15 seconds to destroy the instance on the hypervisor. [ 765.504830] env[61728]: DEBUG oslo.service.loopingcall [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.505910] env[61728]: DEBUG nova.compute.manager [-] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 765.505910] env[61728]: DEBUG nova.network.neutron [-] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 765.551534] env[61728]: DEBUG oslo_vmware.api [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464103, 'name': RemoveSnapshot_Task, 'duration_secs': 0.619226} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.551866] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 765.552197] env[61728]: INFO nova.compute.manager [None req-775d855e-7d22-4d24-ac15-5bf7f7ea4312 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Took 18.69 seconds to snapshot the instance on the hypervisor. [ 765.661941] env[61728]: DEBUG nova.scheduler.client.report [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.730796] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Releasing lock "refresh_cache-b63fa3df-317a-46d4-b8e9-74e9e287efde" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.731170] env[61728]: DEBUG nova.compute.manager [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Instance network_info: |[{"id": "3f4753c4-b820-4467-8f8a-6084733c6e28", "address": "fa:16:3e:ca:57:67", "network": {"id": "6a1908cc-f6aa-4dd7-970e-c8a07e1b16ee", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2047209328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "01839050e56e4dbfa6e22471a3c60998", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f4753c4-b8", "ovs_interfaceid": "3f4753c4-b820-4467-8f8a-6084733c6e28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 765.732800] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:57:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48512b02-ad5c-4105-ba7d-fd4775acf8e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3f4753c4-b820-4467-8f8a-6084733c6e28', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 765.742947] env[61728]: DEBUG oslo.service.loopingcall [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.745115] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 765.745499] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a4293706-1b20-4534-9392-c8c6925ce676 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.766082] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 765.766082] env[61728]: value = "task-464109" [ 765.766082] env[61728]: _type = "Task" [ 765.766082] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.775339] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464109, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.797942] env[61728]: DEBUG oslo_concurrency.lockutils [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] Releasing lock "refresh_cache-39eaa6f4-df3f-4928-a1f1-8c861a392e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.798253] env[61728]: DEBUG nova.compute.manager [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Received event network-vif-plugged-3f4753c4-b820-4467-8f8a-6084733c6e28 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 765.798459] env[61728]: DEBUG oslo_concurrency.lockutils [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] Acquiring lock "b63fa3df-317a-46d4-b8e9-74e9e287efde-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.798691] env[61728]: DEBUG oslo_concurrency.lockutils [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] Lock "b63fa3df-317a-46d4-b8e9-74e9e287efde-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.798862] env[61728]: DEBUG oslo_concurrency.lockutils [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] Lock "b63fa3df-317a-46d4-b8e9-74e9e287efde-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.799042] env[61728]: DEBUG nova.compute.manager [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] No waiting events found dispatching network-vif-plugged-3f4753c4-b820-4467-8f8a-6084733c6e28 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 765.799219] env[61728]: WARNING nova.compute.manager [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Received unexpected event network-vif-plugged-3f4753c4-b820-4467-8f8a-6084733c6e28 for instance with vm_state building and task_state spawning. [ 765.799482] env[61728]: DEBUG nova.compute.manager [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Received event network-changed-3f4753c4-b820-4467-8f8a-6084733c6e28 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 765.799562] env[61728]: DEBUG nova.compute.manager [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Refreshing instance network info cache due to event network-changed-3f4753c4-b820-4467-8f8a-6084733c6e28. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 765.799722] env[61728]: DEBUG oslo_concurrency.lockutils [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] Acquiring lock "refresh_cache-b63fa3df-317a-46d4-b8e9-74e9e287efde" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.799863] env[61728]: DEBUG oslo_concurrency.lockutils [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] Acquired lock "refresh_cache-b63fa3df-317a-46d4-b8e9-74e9e287efde" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.800076] env[61728]: DEBUG nova.network.neutron [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Refreshing network info cache for port 3f4753c4-b820-4467-8f8a-6084733c6e28 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 765.839023] env[61728]: DEBUG oslo_vmware.api [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464108, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147847} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.839023] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 765.839023] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 765.839023] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 765.839023] env[61728]: INFO nova.compute.manager [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Took 1.10 seconds to destroy the instance on the hypervisor. [ 765.839023] env[61728]: DEBUG oslo.service.loopingcall [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.839023] env[61728]: DEBUG nova.compute.manager [-] [instance: 2e83145a-2481-411d-b037-a305a209074e] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 765.839023] env[61728]: DEBUG nova.network.neutron [-] [instance: 2e83145a-2481-411d-b037-a305a209074e] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 765.853514] env[61728]: DEBUG nova.network.neutron [-] [instance: 2e83145a-2481-411d-b037-a305a209074e] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 765.944954] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5252841a-5840-4037-3454-d544fc47b2b1, 'name': SearchDatastore_Task, 'duration_secs': 0.015453} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.945303] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.945734] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 39eaa6f4-df3f-4928-a1f1-8c861a392e29/39eaa6f4-df3f-4928-a1f1-8c861a392e29.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 765.946020] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-754c2005-d48f-452a-98d9-b48413cb257e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.954020] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Waiting for the task: (returnval){ [ 765.954020] env[61728]: value = "task-464110" [ 765.954020] env[61728]: _type = "Task" [ 765.954020] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.962321] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464110, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.168806] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.716s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.169425] env[61728]: DEBUG nova.compute.manager [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 766.172794] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 42.507s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.174545] env[61728]: INFO nova.compute.claims [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.246569] env[61728]: DEBUG nova.network.neutron [-] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.277634] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464109, 'name': CreateVM_Task, 'duration_secs': 0.446524} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.277848] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 766.278625] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.278803] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.279205] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 766.279910] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e3e7688-768a-417b-9b0a-f2a5c9e60603 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.287415] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 766.287415] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bd6f08-4358-d0af-1da1-668a0c2fc4bb" [ 766.287415] env[61728]: _type = "Task" [ 766.287415] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.298060] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bd6f08-4358-d0af-1da1-668a0c2fc4bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.356683] env[61728]: DEBUG nova.network.neutron [-] [instance: 2e83145a-2481-411d-b037-a305a209074e] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.467269] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464110, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487574} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.467540] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 39eaa6f4-df3f-4928-a1f1-8c861a392e29/39eaa6f4-df3f-4928-a1f1-8c861a392e29.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 766.467768] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 766.468018] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-52a84699-b6fa-4c29-97be-02f3c63ab1b1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.475272] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Waiting for the task: (returnval){ [ 766.475272] env[61728]: value = "task-464111" [ 766.475272] env[61728]: _type = "Task" [ 766.475272] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.486287] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464111, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.594912] env[61728]: DEBUG nova.network.neutron [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance_info_cache with network_info: [{"id": "6c9a6b61-5248-4996-a390-c18edb3ec021", "address": "fa:16:3e:43:9a:d5", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9a6b61-52", "ovs_interfaceid": "6c9a6b61-5248-4996-a390-c18edb3ec021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.627582] env[61728]: DEBUG nova.network.neutron [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Updated VIF entry in instance network info cache for port 3f4753c4-b820-4467-8f8a-6084733c6e28. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 766.627582] env[61728]: DEBUG nova.network.neutron [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Updating instance_info_cache with network_info: [{"id": "3f4753c4-b820-4467-8f8a-6084733c6e28", "address": "fa:16:3e:ca:57:67", "network": {"id": "6a1908cc-f6aa-4dd7-970e-c8a07e1b16ee", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2047209328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "01839050e56e4dbfa6e22471a3c60998", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f4753c4-b8", "ovs_interfaceid": "3f4753c4-b820-4467-8f8a-6084733c6e28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.681568] env[61728]: DEBUG nova.compute.utils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 766.683017] env[61728]: DEBUG nova.compute.manager [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 766.683169] env[61728]: DEBUG nova.network.neutron [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 766.724267] env[61728]: DEBUG nova.policy [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '70856f8255c940ac979244efb02c43b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '276942c8cf8a42729a541de096f69a0c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 766.733010] env[61728]: DEBUG nova.compute.manager [req-2cdbe417-6e9d-45d5-b19f-067ea081a847 req-e942c7c3-7d6c-4f54-a356-0916736d3ddc service nova] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Received event network-vif-deleted-d850f13b-cef8-4936-a8a4-f070c7b531e8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 766.742813] env[61728]: DEBUG oslo_concurrency.lockutils [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.743068] env[61728]: DEBUG oslo_concurrency.lockutils [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.749923] env[61728]: INFO nova.compute.manager [-] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Took 1.24 seconds to deallocate network for instance. [ 766.797976] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bd6f08-4358-d0af-1da1-668a0c2fc4bb, 'name': SearchDatastore_Task, 'duration_secs': 0.056856} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.798307] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.798543] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 766.798805] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.798952] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.799153] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 766.799416] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b444a0e3-0d96-4715-9d76-bea28cfe49a4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.808432] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 766.809232] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 766.809359] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcab12f1-5da4-441b-8f21-f5758f6d5e35 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.815674] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 766.815674] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52154517-7ba6-94f4-1876-b95e33872a48" [ 766.815674] env[61728]: _type = "Task" [ 766.815674] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.824457] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52154517-7ba6-94f4-1876-b95e33872a48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.859679] env[61728]: INFO nova.compute.manager [-] [instance: 2e83145a-2481-411d-b037-a305a209074e] Took 1.02 seconds to deallocate network for instance. [ 766.986731] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464111, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077541} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.987140] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 766.988158] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a71ac2-7d97-4330-ba71-b790dd7afba1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.014335] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 39eaa6f4-df3f-4928-a1f1-8c861a392e29/39eaa6f4-df3f-4928-a1f1-8c861a392e29.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 767.015067] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd660e28-756c-46d1-af30-59594ed47564 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.031028] env[61728]: DEBUG nova.network.neutron [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Successfully created port: 33a9ae57-3181-42d6-9a7f-a66bf8103138 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 767.040324] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Waiting for the task: (returnval){ [ 767.040324] env[61728]: value = "task-464112" [ 767.040324] env[61728]: _type = "Task" [ 767.040324] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.049150] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464112, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.097995] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.098371] env[61728]: DEBUG nova.objects.instance [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'migration_context' on Instance uuid 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 767.130284] env[61728]: DEBUG oslo_concurrency.lockutils [req-5de188ac-0013-4a4f-bf0d-6434a55b541a req-3346bb65-9be3-4088-ba5d-744de30517b2 service nova] Releasing lock "refresh_cache-b63fa3df-317a-46d4-b8e9-74e9e287efde" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.187182] env[61728]: DEBUG nova.compute.manager [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 767.246390] env[61728]: DEBUG nova.compute.utils [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 767.261038] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.326633] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52154517-7ba6-94f4-1876-b95e33872a48, 'name': SearchDatastore_Task, 'duration_secs': 0.009564} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.330310] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb4ec954-e66f-47dd-a6ae-87e416f551db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.335645] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 767.335645] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5247170c-13e9-9efa-1069-0964c659770b" [ 767.335645] env[61728]: _type = "Task" [ 767.335645] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.346195] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5247170c-13e9-9efa-1069-0964c659770b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.366473] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.552126] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464112, 'name': ReconfigVM_Task, 'duration_secs': 0.326325} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.554765] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 39eaa6f4-df3f-4928-a1f1-8c861a392e29/39eaa6f4-df3f-4928-a1f1-8c861a392e29.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 767.556042] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-16969695-c83e-459a-a982-2d1a0dd6d852 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.562869] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Waiting for the task: (returnval){ [ 767.562869] env[61728]: value = "task-464113" [ 767.562869] env[61728]: _type = "Task" [ 767.562869] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.572488] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464113, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.602011] env[61728]: DEBUG nova.objects.base [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Object Instance<7b16fbc0-7f13-405f-b84e-e18de1ca7dd2> lazy-loaded attributes: info_cache,migration_context {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 767.603220] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cccca4f-e7c9-4c8a-83d9-9e7772c0ac29 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.627552] env[61728]: DEBUG nova.compute.manager [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.631838] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1c2e08-9e13-43fc-8ab0-c8c84e2013d9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.634213] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6028c706-ca53-467b-ac5e-7c54487f81e5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.640595] env[61728]: DEBUG oslo_vmware.api [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 767.640595] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5286606b-7190-4d23-2690-bf327e3b49be" [ 767.640595] env[61728]: _type = "Task" [ 767.640595] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.654021] env[61728]: DEBUG oslo_vmware.api [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5286606b-7190-4d23-2690-bf327e3b49be, 'name': SearchDatastore_Task, 'duration_secs': 0.007502} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.654734] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.689542] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa916d6-63ec-4d75-bf64-d5cb015967ef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.703009] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f404e6-aec6-4d4d-bee8-46db521da3ac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.737144] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dee0c6e-d00c-48f2-acb1-fe8ea5a0398b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.746014] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1f3d50-18b4-4efd-822a-fbcfe096b940 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.750523] env[61728]: DEBUG oslo_concurrency.lockutils [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.761901] env[61728]: DEBUG nova.compute.provider_tree [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.846388] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5247170c-13e9-9efa-1069-0964c659770b, 'name': SearchDatastore_Task, 'duration_secs': 0.009747} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.846761] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.847062] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b63fa3df-317a-46d4-b8e9-74e9e287efde/b63fa3df-317a-46d4-b8e9-74e9e287efde.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 767.847416] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-53c2b04b-1503-4499-b8ac-085c2c8a78f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.855716] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 767.855716] env[61728]: value = "task-464114" [ 767.855716] env[61728]: _type = "Task" [ 767.855716] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.865449] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464114, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.076602] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464113, 'name': Rename_Task, 'duration_secs': 0.179615} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.076932] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 768.077270] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d5d2ab4a-c29e-4d40-95a5-fa955c52ea1c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.089186] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Waiting for the task: (returnval){ [ 768.089186] env[61728]: value = "task-464115" [ 768.089186] env[61728]: _type = "Task" [ 768.089186] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.099301] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464115, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.149105] env[61728]: INFO nova.compute.manager [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] instance snapshotting [ 768.152191] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f509bad8-4009-4623-876f-a2c3009ded3a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.177578] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08bcdc84-ac76-4ced-9321-343f22acb287 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.198985] env[61728]: DEBUG nova.compute.manager [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 768.226485] env[61728]: DEBUG nova.virt.hardware [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 768.226929] env[61728]: DEBUG nova.virt.hardware [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 768.227199] env[61728]: DEBUG nova.virt.hardware [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 768.227518] env[61728]: DEBUG nova.virt.hardware [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 768.227781] env[61728]: DEBUG nova.virt.hardware [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 768.228011] env[61728]: DEBUG nova.virt.hardware [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 768.228277] env[61728]: DEBUG nova.virt.hardware [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 768.228533] env[61728]: DEBUG nova.virt.hardware [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 768.228763] env[61728]: DEBUG nova.virt.hardware [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 768.228975] env[61728]: DEBUG nova.virt.hardware [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 768.229231] env[61728]: DEBUG nova.virt.hardware [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 768.230556] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a670ee-b758-401c-be3e-0ee239101392 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.242178] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0cebc1-0ccf-4655-a0af-88dede6b5254 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.266027] env[61728]: DEBUG nova.scheduler.client.report [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.367242] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464114, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485296} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.367539] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b63fa3df-317a-46d4-b8e9-74e9e287efde/b63fa3df-317a-46d4-b8e9-74e9e287efde.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 768.367840] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 768.368136] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-088cfb7e-368a-4c17-bb37-c310b129728e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.377875] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 768.377875] env[61728]: value = "task-464116" [ 768.377875] env[61728]: _type = "Task" [ 768.377875] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.388322] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464116, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.540847] env[61728]: DEBUG nova.network.neutron [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Successfully updated port: 33a9ae57-3181-42d6-9a7f-a66bf8103138 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 768.600583] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464115, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.691182] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 768.691543] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f532829e-86c6-429b-9339-9b3d735a9adc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.700071] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 768.700071] env[61728]: value = "task-464117" [ 768.700071] env[61728]: _type = "Task" [ 768.700071] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.709391] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464117, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.774018] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.599s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.774018] env[61728]: DEBUG nova.compute.manager [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 768.775632] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 44.595s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.776169] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.782057] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61728) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 768.782057] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.437s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.782057] env[61728]: DEBUG nova.objects.instance [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lazy-loading 'resources' on Instance uuid 223b1f93-d33e-4566-8c8e-75234a491814 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 768.783618] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c497c8-a003-4e86-93a9-bde6069cc1b7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.793766] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88c63ec3-cc76-489d-9673-f1b8811dd621 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.810907] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d00b53-c7c9-4818-91be-33e07ccc5b4f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.819891] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a392a89b-affb-4ca8-b8b6-0d97271efe00 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.851870] env[61728]: DEBUG oslo_concurrency.lockutils [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.852159] env[61728]: DEBUG oslo_concurrency.lockutils [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.852411] env[61728]: INFO nova.compute.manager [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Attaching volume 11c73a47-a9c0-4b47-a431-15f14807f4f9 to /dev/sdb [ 768.854801] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178335MB free_disk=112GB free_vcpus=48 pci_devices=None {{(pid=61728) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 768.854940] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.870481] env[61728]: DEBUG nova.compute.manager [req-b948b1a7-6873-4a49-96a1-f87e8dda27b1 req-a50ced99-ca9a-4782-a315-1dfa62967637 service nova] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Received event network-vif-plugged-33a9ae57-3181-42d6-9a7f-a66bf8103138 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 768.870729] env[61728]: DEBUG oslo_concurrency.lockutils [req-b948b1a7-6873-4a49-96a1-f87e8dda27b1 req-a50ced99-ca9a-4782-a315-1dfa62967637 service nova] Acquiring lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.870993] env[61728]: DEBUG oslo_concurrency.lockutils [req-b948b1a7-6873-4a49-96a1-f87e8dda27b1 req-a50ced99-ca9a-4782-a315-1dfa62967637 service nova] Lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.871210] env[61728]: DEBUG oslo_concurrency.lockutils [req-b948b1a7-6873-4a49-96a1-f87e8dda27b1 req-a50ced99-ca9a-4782-a315-1dfa62967637 service nova] Lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.871325] env[61728]: DEBUG nova.compute.manager [req-b948b1a7-6873-4a49-96a1-f87e8dda27b1 req-a50ced99-ca9a-4782-a315-1dfa62967637 service nova] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] No waiting events found dispatching network-vif-plugged-33a9ae57-3181-42d6-9a7f-a66bf8103138 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 768.871505] env[61728]: WARNING nova.compute.manager [req-b948b1a7-6873-4a49-96a1-f87e8dda27b1 req-a50ced99-ca9a-4782-a315-1dfa62967637 service nova] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Received unexpected event network-vif-plugged-33a9ae57-3181-42d6-9a7f-a66bf8103138 for instance with vm_state building and task_state spawning. [ 768.871711] env[61728]: DEBUG nova.compute.manager [req-b948b1a7-6873-4a49-96a1-f87e8dda27b1 req-a50ced99-ca9a-4782-a315-1dfa62967637 service nova] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Received event network-changed-33a9ae57-3181-42d6-9a7f-a66bf8103138 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 768.871883] env[61728]: DEBUG nova.compute.manager [req-b948b1a7-6873-4a49-96a1-f87e8dda27b1 req-a50ced99-ca9a-4782-a315-1dfa62967637 service nova] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Refreshing instance network info cache due to event network-changed-33a9ae57-3181-42d6-9a7f-a66bf8103138. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 768.872100] env[61728]: DEBUG oslo_concurrency.lockutils [req-b948b1a7-6873-4a49-96a1-f87e8dda27b1 req-a50ced99-ca9a-4782-a315-1dfa62967637 service nova] Acquiring lock "refresh_cache-06e765e1-85d9-4f6c-8ab5-81735f08cb66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.872244] env[61728]: DEBUG oslo_concurrency.lockutils [req-b948b1a7-6873-4a49-96a1-f87e8dda27b1 req-a50ced99-ca9a-4782-a315-1dfa62967637 service nova] Acquired lock "refresh_cache-06e765e1-85d9-4f6c-8ab5-81735f08cb66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.872416] env[61728]: DEBUG nova.network.neutron [req-b948b1a7-6873-4a49-96a1-f87e8dda27b1 req-a50ced99-ca9a-4782-a315-1dfa62967637 service nova] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Refreshing network info cache for port 33a9ae57-3181-42d6-9a7f-a66bf8103138 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 768.886732] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63473f39-ee92-4c44-bf20-391c34720081 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.893040] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464116, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086949} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.893699] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 768.894483] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-791684d4-35b1-4303-88d8-427f508e9ac7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.899358] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeae9fcb-6598-4239-9af8-7abe25fe8c29 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.919354] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] b63fa3df-317a-46d4-b8e9-74e9e287efde/b63fa3df-317a-46d4-b8e9-74e9e287efde.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 768.920191] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27401b3e-f62f-41f1-ba08-29ec833c660d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.938790] env[61728]: DEBUG nova.virt.block_device [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Updating existing volume attachment record: c820edf6-c15b-4827-89d2-d0292f807ebd {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 768.947587] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 768.947587] env[61728]: value = "task-464118" [ 768.947587] env[61728]: _type = "Task" [ 768.947587] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.956315] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464118, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.043885] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "refresh_cache-06e765e1-85d9-4f6c-8ab5-81735f08cb66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.110165] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464115, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.210685] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464117, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.287183] env[61728]: DEBUG nova.compute.utils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 769.288734] env[61728]: DEBUG nova.compute.manager [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 769.288908] env[61728]: DEBUG nova.network.neutron [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 769.332071] env[61728]: DEBUG nova.policy [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f153875691ce4623b3bebb198a1d3e7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e0776c511b94f2e88ded3b52f5a7b76', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 769.436790] env[61728]: DEBUG nova.network.neutron [req-b948b1a7-6873-4a49-96a1-f87e8dda27b1 req-a50ced99-ca9a-4782-a315-1dfa62967637 service nova] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 769.460844] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464118, 'name': ReconfigVM_Task, 'duration_secs': 0.432504} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.463651] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Reconfigured VM instance instance-0000003a to attach disk [datastore1] b63fa3df-317a-46d4-b8e9-74e9e287efde/b63fa3df-317a-46d4-b8e9-74e9e287efde.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 769.465616] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3c8d0f4d-c6b3-4d2a-929a-ca75e6ba33f6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.475645] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 769.475645] env[61728]: value = "task-464122" [ 769.475645] env[61728]: _type = "Task" [ 769.475645] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.493059] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464122, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.585720] env[61728]: DEBUG nova.network.neutron [req-b948b1a7-6873-4a49-96a1-f87e8dda27b1 req-a50ced99-ca9a-4782-a315-1dfa62967637 service nova] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.607116] env[61728]: DEBUG oslo_vmware.api [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464115, 'name': PowerOnVM_Task, 'duration_secs': 1.126234} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.609753] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 769.609998] env[61728]: INFO nova.compute.manager [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Took 8.01 seconds to spawn the instance on the hypervisor. [ 769.610224] env[61728]: DEBUG nova.compute.manager [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 769.611548] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feddf024-734e-4d1d-bad0-2f590624fbfa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.717695] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464117, 'name': CreateSnapshot_Task, 'duration_secs': 0.642854} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.718027] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 769.718839] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84523fb2-3fae-48ac-b91c-533b2650189d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.755739] env[61728]: DEBUG nova.network.neutron [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Successfully created port: 6f900c64-beea-4e8a-84c5-9f13768f8420 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.791738] env[61728]: DEBUG nova.compute.manager [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.929746] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249de26a-5b1c-434f-9032-41ec844b4037 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.938109] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3bbfea-8c78-4c06-8c8f-306fbb947237 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.974025] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add008a3-576c-4074-b6f3-2979aee6dd6d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.987067] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464122, 'name': Rename_Task, 'duration_secs': 0.199608} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.987449] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 769.988651] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4ffc13-22de-4a7a-9a05-d0093431ad84 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.992531] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2fed48a5-edca-4f1f-9e5b-336d6bb9ba07 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.005456] env[61728]: DEBUG nova.compute.provider_tree [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.007561] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 770.007561] env[61728]: value = "task-464123" [ 770.007561] env[61728]: _type = "Task" [ 770.007561] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.016780] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464123, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.088354] env[61728]: DEBUG oslo_concurrency.lockutils [req-b948b1a7-6873-4a49-96a1-f87e8dda27b1 req-a50ced99-ca9a-4782-a315-1dfa62967637 service nova] Releasing lock "refresh_cache-06e765e1-85d9-4f6c-8ab5-81735f08cb66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.088775] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired lock "refresh_cache-06e765e1-85d9-4f6c-8ab5-81735f08cb66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.088986] env[61728]: DEBUG nova.network.neutron [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 770.132538] env[61728]: INFO nova.compute.manager [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Took 55.26 seconds to build instance. [ 770.247017] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 770.248234] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-af1a1932-a206-4cda-966d-8b766f044227 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.257448] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 770.257448] env[61728]: value = "task-464124" [ 770.257448] env[61728]: _type = "Task" [ 770.257448] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.266522] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464124, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.509206] env[61728]: DEBUG nova.scheduler.client.report [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.522830] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464123, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.620066] env[61728]: DEBUG nova.network.neutron [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.635446] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d183c6d6-33d9-4b87-a575-7dd28361fb7d tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Lock "39eaa6f4-df3f-4928-a1f1-8c861a392e29" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.777s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.769489] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464124, 'name': CloneVM_Task} progress is 93%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.773430] env[61728]: DEBUG nova.network.neutron [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Updating instance_info_cache with network_info: [{"id": "33a9ae57-3181-42d6-9a7f-a66bf8103138", "address": "fa:16:3e:9c:55:7f", "network": {"id": "22ad4b68-2895-4651-803f-44b8031a6292", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-489227259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "276942c8cf8a42729a541de096f69a0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33a9ae57-31", "ovs_interfaceid": "33a9ae57-3181-42d6-9a7f-a66bf8103138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.777137] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Acquiring lock "37b8ca5a-e0f8-414b-a363-e56db520f027" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.777398] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Lock "37b8ca5a-e0f8-414b-a363-e56db520f027" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.777582] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Acquiring lock "37b8ca5a-e0f8-414b-a363-e56db520f027-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.777853] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Lock "37b8ca5a-e0f8-414b-a363-e56db520f027-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.778104] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Lock "37b8ca5a-e0f8-414b-a363-e56db520f027-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.780064] env[61728]: INFO nova.compute.manager [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Terminating instance [ 770.782234] env[61728]: DEBUG nova.compute.manager [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 770.782413] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 770.782495] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cd77bfcb-b907-448b-9257-3e66a3485fff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.793840] env[61728]: DEBUG oslo_vmware.api [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Waiting for the task: (returnval){ [ 770.793840] env[61728]: value = "task-464125" [ 770.793840] env[61728]: _type = "Task" [ 770.793840] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.803301] env[61728]: DEBUG nova.compute.manager [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.809902] env[61728]: DEBUG oslo_vmware.api [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-464125, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.834347] env[61728]: DEBUG nova.virt.hardware [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.834622] env[61728]: DEBUG nova.virt.hardware [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.834766] env[61728]: DEBUG nova.virt.hardware [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.834948] env[61728]: DEBUG nova.virt.hardware [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.835112] env[61728]: DEBUG nova.virt.hardware [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.835264] env[61728]: DEBUG nova.virt.hardware [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.835469] env[61728]: DEBUG nova.virt.hardware [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.835625] env[61728]: DEBUG nova.virt.hardware [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.835791] env[61728]: DEBUG nova.virt.hardware [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.835950] env[61728]: DEBUG nova.virt.hardware [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.836135] env[61728]: DEBUG nova.virt.hardware [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.837011] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c86dfd2-773c-40e7-ba4b-14c8f3305273 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.846220] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11764d8-4372-4a70-9287-f4f4c6d28211 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.017822] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.237s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.021087] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 43.416s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.022944] env[61728]: INFO nova.compute.claims [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.039250] env[61728]: DEBUG oslo_vmware.api [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464123, 'name': PowerOnVM_Task, 'duration_secs': 0.998631} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.039625] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 771.039845] env[61728]: INFO nova.compute.manager [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Took 7.83 seconds to spawn the instance on the hypervisor. [ 771.040041] env[61728]: DEBUG nova.compute.manager [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 771.041840] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a97b44-ab8d-4a25-8f9c-7a92cda2d979 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.049562] env[61728]: INFO nova.scheduler.client.report [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Deleted allocations for instance 223b1f93-d33e-4566-8c8e-75234a491814 [ 771.140706] env[61728]: DEBUG nova.compute.manager [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 771.254732] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Acquiring lock "39eaa6f4-df3f-4928-a1f1-8c861a392e29" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.255051] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Lock "39eaa6f4-df3f-4928-a1f1-8c861a392e29" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.255211] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Acquiring lock "39eaa6f4-df3f-4928-a1f1-8c861a392e29-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.255433] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Lock "39eaa6f4-df3f-4928-a1f1-8c861a392e29-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.255656] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Lock "39eaa6f4-df3f-4928-a1f1-8c861a392e29-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.258460] env[61728]: INFO nova.compute.manager [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Terminating instance [ 771.264098] env[61728]: DEBUG nova.compute.manager [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 771.264163] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 771.265251] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf222d1-af60-4890-af29-3f50740cd6c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.275199] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Releasing lock "refresh_cache-06e765e1-85d9-4f6c-8ab5-81735f08cb66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.275509] env[61728]: DEBUG nova.compute.manager [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Instance network_info: |[{"id": "33a9ae57-3181-42d6-9a7f-a66bf8103138", "address": "fa:16:3e:9c:55:7f", "network": {"id": "22ad4b68-2895-4651-803f-44b8031a6292", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-489227259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "276942c8cf8a42729a541de096f69a0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33a9ae57-31", "ovs_interfaceid": "33a9ae57-3181-42d6-9a7f-a66bf8103138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 771.280884] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:55:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0248a27a-1d7f-4195-987b-06bfc8467347', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '33a9ae57-3181-42d6-9a7f-a66bf8103138', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 771.290410] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Creating folder: Project (276942c8cf8a42729a541de096f69a0c). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 771.290780] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 771.291023] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464124, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.291268] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b37eef32-22d2-4abc-ba9c-91f7317741cb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.293124] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54761e6d-9329-4dc9-8d90-bdadb863a27c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.306535] env[61728]: DEBUG oslo_vmware.api [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-464125, 'name': PowerOffVM_Task, 'duration_secs': 0.271813} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.309199] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 771.309372] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 771.309622] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-121945', 'volume_id': '2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'name': 'volume-2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '37b8ca5a-e0f8-414b-a363-e56db520f027', 'attached_at': '', 'detached_at': '', 'volume_id': '2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'serial': '2e85ae80-363c-4fa7-add2-31dd95e23cfd'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 771.309994] env[61728]: DEBUG oslo_vmware.api [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Waiting for the task: (returnval){ [ 771.309994] env[61728]: value = "task-464127" [ 771.309994] env[61728]: _type = "Task" [ 771.309994] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.310235] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Created folder: Project (276942c8cf8a42729a541de096f69a0c) in parent group-v121913. [ 771.310441] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Creating folder: Instances. Parent ref: group-v122094. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 771.311267] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c0ec4cf-a828-4dc4-9254-3fcf8816808c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.314338] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fe5b402c-598a-454d-aa25-77777a628cd8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.332359] env[61728]: DEBUG oslo_vmware.api [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464127, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.349248] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e83ce3-42f8-43e9-a82d-5b811e68934a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.352417] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Created folder: Instances in parent group-v122094. [ 771.352709] env[61728]: DEBUG oslo.service.loopingcall [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.352960] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 771.353662] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-59f10d35-9703-43a0-b00c-efda76308d9c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.370410] env[61728]: DEBUG nova.network.neutron [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Successfully updated port: 6f900c64-beea-4e8a-84c5-9f13768f8420 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 771.374943] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca8dcc5c-ea24-4766-a3c1-4b8db834e653 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.380320] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 771.380320] env[61728]: value = "task-464129" [ 771.380320] env[61728]: _type = "Task" [ 771.380320] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.398881] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d1f463-0975-41e6-9077-faad3fb14e32 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.405520] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464129, 'name': CreateVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.421514] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] The volume has not been displaced from its original location: [datastore1] volume-2e85ae80-363c-4fa7-add2-31dd95e23cfd/volume-2e85ae80-363c-4fa7-add2-31dd95e23cfd.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 771.427214] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Reconfiguring VM instance instance-0000001a to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 771.427626] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5877f21-af67-478b-b0a8-e7e57ca62b86 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.448613] env[61728]: DEBUG oslo_vmware.api [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Waiting for the task: (returnval){ [ 771.448613] env[61728]: value = "task-464130" [ 771.448613] env[61728]: _type = "Task" [ 771.448613] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.460541] env[61728]: DEBUG oslo_vmware.api [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-464130, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.574586] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f70cba93-910b-40f9-97bb-93a9eccf8518 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "223b1f93-d33e-4566-8c8e-75234a491814" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 47.221s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.578338] env[61728]: INFO nova.compute.manager [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Took 53.82 seconds to build instance. [ 771.585536] env[61728]: DEBUG nova.compute.manager [req-72c56c33-5d16-4bea-be17-89207d556a2e req-2d4bc949-cc7f-4091-846e-a15d80cb945a service nova] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Received event network-vif-plugged-6f900c64-beea-4e8a-84c5-9f13768f8420 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 771.586093] env[61728]: DEBUG oslo_concurrency.lockutils [req-72c56c33-5d16-4bea-be17-89207d556a2e req-2d4bc949-cc7f-4091-846e-a15d80cb945a service nova] Acquiring lock "a5c7e3ff-6829-4528-9279-e865a0eb8512-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.586375] env[61728]: DEBUG oslo_concurrency.lockutils [req-72c56c33-5d16-4bea-be17-89207d556a2e req-2d4bc949-cc7f-4091-846e-a15d80cb945a service nova] Lock "a5c7e3ff-6829-4528-9279-e865a0eb8512-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.586375] env[61728]: DEBUG oslo_concurrency.lockutils [req-72c56c33-5d16-4bea-be17-89207d556a2e req-2d4bc949-cc7f-4091-846e-a15d80cb945a service nova] Lock "a5c7e3ff-6829-4528-9279-e865a0eb8512-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.586510] env[61728]: DEBUG nova.compute.manager [req-72c56c33-5d16-4bea-be17-89207d556a2e req-2d4bc949-cc7f-4091-846e-a15d80cb945a service nova] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] No waiting events found dispatching network-vif-plugged-6f900c64-beea-4e8a-84c5-9f13768f8420 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 771.586691] env[61728]: WARNING nova.compute.manager [req-72c56c33-5d16-4bea-be17-89207d556a2e req-2d4bc949-cc7f-4091-846e-a15d80cb945a service nova] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Received unexpected event network-vif-plugged-6f900c64-beea-4e8a-84c5-9f13768f8420 for instance with vm_state building and task_state spawning. [ 771.664130] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.775814] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464124, 'name': CloneVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.828668] env[61728]: DEBUG oslo_vmware.api [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464127, 'name': PowerOffVM_Task, 'duration_secs': 0.21068} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.829233] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 771.829652] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 771.830382] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-181f1446-6f56-4f40-bc4c-808a75d440f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.874440] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Acquiring lock "refresh_cache-a5c7e3ff-6829-4528-9279-e865a0eb8512" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.874503] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Acquired lock "refresh_cache-a5c7e3ff-6829-4528-9279-e865a0eb8512" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.874793] env[61728]: DEBUG nova.network.neutron [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 771.892891] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464129, 'name': CreateVM_Task, 'duration_secs': 0.39224} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.894108] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 771.895219] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.895544] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.896131] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 771.896834] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4c50ac0-1987-4795-a494-44bdb230c64b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.905989] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 771.905989] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52720584-42f2-3485-23f8-71ff4656e49d" [ 771.905989] env[61728]: _type = "Task" [ 771.905989] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.922200] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52720584-42f2-3485-23f8-71ff4656e49d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.924102] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 771.924447] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 771.924755] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Deleting the datastore file [datastore1] 39eaa6f4-df3f-4928-a1f1-8c861a392e29 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 771.925152] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1aba18e9-5863-4048-a200-d58f976c7a08 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.934684] env[61728]: DEBUG oslo_vmware.api [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Waiting for the task: (returnval){ [ 771.934684] env[61728]: value = "task-464133" [ 771.934684] env[61728]: _type = "Task" [ 771.934684] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.946759] env[61728]: DEBUG oslo_vmware.api [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464133, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.958393] env[61728]: DEBUG oslo_vmware.api [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-464130, 'name': ReconfigVM_Task, 'duration_secs': 0.211222} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.958677] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Reconfigured VM instance instance-0000001a to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 771.963883] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-324e3768-44c5-49de-a540-bc4a7791f7ee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.981558] env[61728]: DEBUG oslo_vmware.api [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Waiting for the task: (returnval){ [ 771.981558] env[61728]: value = "task-464134" [ 771.981558] env[61728]: _type = "Task" [ 771.981558] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.991398] env[61728]: DEBUG oslo_vmware.api [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-464134, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.083709] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b67c854b-21d1-456b-906b-1c8fc29f0287 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "b63fa3df-317a-46d4-b8e9-74e9e287efde" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.207s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.275539] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464124, 'name': CloneVM_Task, 'duration_secs': 1.634841} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.275872] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Created linked-clone VM from snapshot [ 772.276592] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f64468-4f74-45ef-93a2-1d1271c1044e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.287369] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Uploading image 05e97342-0afd-4bd8-b341-7f3386bdf448 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 772.310341] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 772.310341] env[61728]: value = "vm-122093" [ 772.310341] env[61728]: _type = "VirtualMachine" [ 772.310341] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 772.310803] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-23368197-317e-4eb3-b18b-2c645e60a261 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.320212] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lease: (returnval){ [ 772.320212] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528bcbfe-099f-21b2-7397-d8e09976c887" [ 772.320212] env[61728]: _type = "HttpNfcLease" [ 772.320212] env[61728]: } obtained for exporting VM: (result){ [ 772.320212] env[61728]: value = "vm-122093" [ 772.320212] env[61728]: _type = "VirtualMachine" [ 772.320212] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 772.320494] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the lease: (returnval){ [ 772.320494] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528bcbfe-099f-21b2-7397-d8e09976c887" [ 772.320494] env[61728]: _type = "HttpNfcLease" [ 772.320494] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 772.331107] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 772.331107] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528bcbfe-099f-21b2-7397-d8e09976c887" [ 772.331107] env[61728]: _type = "HttpNfcLease" [ 772.331107] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 772.405794] env[61728]: DEBUG nova.network.neutron [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 772.417668] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52720584-42f2-3485-23f8-71ff4656e49d, 'name': SearchDatastore_Task, 'duration_secs': 0.019157} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.418106] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.418362] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.418605] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.419186] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.419437] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.419823] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d057f5e3-d8ab-49cd-9530-2976d1d5d362 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.434749] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 772.434969] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 772.435803] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e913f5d-1f4c-4756-a28c-cfa19e163996 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.452024] env[61728]: DEBUG oslo_vmware.api [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464133, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.454124] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 772.454124] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b56220-666f-a5a2-0c50-04ea6d2a4a3a" [ 772.454124] env[61728]: _type = "Task" [ 772.454124] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.463451] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b56220-666f-a5a2-0c50-04ea6d2a4a3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.492299] env[61728]: DEBUG oslo_vmware.api [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-464134, 'name': ReconfigVM_Task, 'duration_secs': 0.169405} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.495231] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-121945', 'volume_id': '2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'name': 'volume-2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '37b8ca5a-e0f8-414b-a363-e56db520f027', 'attached_at': '', 'detached_at': '', 'volume_id': '2e85ae80-363c-4fa7-add2-31dd95e23cfd', 'serial': '2e85ae80-363c-4fa7-add2-31dd95e23cfd'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 772.495583] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 772.496639] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b38ee0ce-d57b-4748-a5e8-40008f0dcefa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.504893] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 772.507706] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1fb5883d-5bad-4a50-b60d-2853faf15304 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.589861] env[61728]: DEBUG nova.compute.manager [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 772.599682] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 772.605027] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 772.605027] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Deleting the datastore file [datastore1] 37b8ca5a-e0f8-414b-a363-e56db520f027 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 772.605027] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d2d805da-d887-4e10-a70d-d8d360178fe9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.611842] env[61728]: DEBUG nova.network.neutron [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Updating instance_info_cache with network_info: [{"id": "6f900c64-beea-4e8a-84c5-9f13768f8420", "address": "fa:16:3e:3c:d1:e3", "network": {"id": "4d4f744b-c837-4ae8-9d82-7dce6a34b344", "bridge": "br-int", "label": "tempest-ServersTestJSON-1248642690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0776c511b94f2e88ded3b52f5a7b76", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f900c64-be", "ovs_interfaceid": "6f900c64-beea-4e8a-84c5-9f13768f8420", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.616320] env[61728]: DEBUG oslo_vmware.api [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Waiting for the task: (returnval){ [ 772.616320] env[61728]: value = "task-464137" [ 772.616320] env[61728]: _type = "Task" [ 772.616320] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.635309] env[61728]: DEBUG oslo_vmware.api [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-464137, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.654157] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a48f20-a388-47e5-8fd7-c4111854769e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.664280] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26147530-7bd2-4efe-a159-eae8eee223be {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.696090] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f33fa302-b4e8-431a-8aa2-b2f07f4db228 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.706141] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72cd0f9-e849-4c86-abcf-f4b482bc3bfc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.721739] env[61728]: DEBUG nova.compute.provider_tree [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.832080] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 772.832080] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528bcbfe-099f-21b2-7397-d8e09976c887" [ 772.832080] env[61728]: _type = "HttpNfcLease" [ 772.832080] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 772.832080] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 772.832080] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528bcbfe-099f-21b2-7397-d8e09976c887" [ 772.832080] env[61728]: _type = "HttpNfcLease" [ 772.832080] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 772.832850] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ced1732-4d5a-40f4-9fc1-4081de7c60ce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.841985] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d76f18-e78c-3017-7ef0-ca2c35a87c0e/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 772.842132] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d76f18-e78c-3017-7ef0-ca2c35a87c0e/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 772.942583] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d834cbaa-5d1e-42e7-a3cb-49290642baa9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.951427] env[61728]: DEBUG oslo_vmware.api [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Task: {'id': task-464133, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.676681} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.951831] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 772.952070] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 772.952293] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 772.952481] env[61728]: INFO nova.compute.manager [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Took 1.69 seconds to destroy the instance on the hypervisor. [ 772.952744] env[61728]: DEBUG oslo.service.loopingcall [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.952858] env[61728]: DEBUG nova.compute.manager [-] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 772.952957] env[61728]: DEBUG nova.network.neutron [-] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 772.969898] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b56220-666f-a5a2-0c50-04ea6d2a4a3a, 'name': SearchDatastore_Task, 'duration_secs': 0.026515} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.973986] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e84b181-0c62-4075-ba66-2b095c854594 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.986328] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 772.986328] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e64ad3-5b30-f332-ad86-5c1eef007bbd" [ 772.986328] env[61728]: _type = "Task" [ 772.986328] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.995638] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e64ad3-5b30-f332-ad86-5c1eef007bbd, 'name': SearchDatastore_Task, 'duration_secs': 0.011885} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.995893] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.996167] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 06e765e1-85d9-4f6c-8ab5-81735f08cb66/06e765e1-85d9-4f6c-8ab5-81735f08cb66.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 772.996433] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-85f76b88-c3c3-4162-990b-db990f50fd9a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.004931] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 773.004931] env[61728]: value = "task-464138" [ 773.004931] env[61728]: _type = "Task" [ 773.004931] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.014030] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464138, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.119144] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Releasing lock "refresh_cache-a5c7e3ff-6829-4528-9279-e865a0eb8512" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.119144] env[61728]: DEBUG nova.compute.manager [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Instance network_info: |[{"id": "6f900c64-beea-4e8a-84c5-9f13768f8420", "address": "fa:16:3e:3c:d1:e3", "network": {"id": "4d4f744b-c837-4ae8-9d82-7dce6a34b344", "bridge": "br-int", "label": "tempest-ServersTestJSON-1248642690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0776c511b94f2e88ded3b52f5a7b76", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f900c64-be", "ovs_interfaceid": "6f900c64-beea-4e8a-84c5-9f13768f8420", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 773.119144] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:d1:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6f900c64-beea-4e8a-84c5-9f13768f8420', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 773.127920] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Creating folder: Project (3e0776c511b94f2e88ded3b52f5a7b76). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 773.134878] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.134878] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6bc57221-ef99-4ed6-9a1f-b847407ff62e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.144442] env[61728]: DEBUG oslo_vmware.api [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Task: {'id': task-464137, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.301429} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.144799] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 773.144994] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 773.145201] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 773.145381] env[61728]: INFO nova.compute.manager [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Took 2.36 seconds to destroy the instance on the hypervisor. [ 773.145643] env[61728]: DEBUG oslo.service.loopingcall [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 773.145870] env[61728]: DEBUG nova.compute.manager [-] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 773.145965] env[61728]: DEBUG nova.network.neutron [-] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 773.149690] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Created folder: Project (3e0776c511b94f2e88ded3b52f5a7b76) in parent group-v121913. [ 773.150208] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Creating folder: Instances. Parent ref: group-v122097. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 773.151409] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c723b43f-354c-4a9a-ae63-1aa7ce1d03be {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.164759] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Created folder: Instances in parent group-v122097. [ 773.164859] env[61728]: DEBUG oslo.service.loopingcall [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 773.165070] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 773.165354] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ed2920d-86e7-4367-8e53-ee5383ffb768 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.196018] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 773.196018] env[61728]: value = "task-464141" [ 773.196018] env[61728]: _type = "Task" [ 773.196018] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.207418] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464141, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.225758] env[61728]: DEBUG nova.scheduler.client.report [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.521090] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464138, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450159} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.521328] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 06e765e1-85d9-4f6c-8ab5-81735f08cb66/06e765e1-85d9-4f6c-8ab5-81735f08cb66.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 773.524903] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 773.525372] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19537367-41fe-4847-a8e8-6ad07e69801e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.541122] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 773.541122] env[61728]: value = "task-464142" [ 773.541122] env[61728]: _type = "Task" [ 773.541122] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.558785] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464142, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.710902] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464141, 'name': CreateVM_Task, 'duration_secs': 0.431076} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.711038] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 773.711817] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.711999] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.712357] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 773.712807] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-961072b8-0129-4611-9118-ed9564fe8344 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.718904] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Waiting for the task: (returnval){ [ 773.718904] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522ec186-a660-865e-6b6f-45e96c32aef2" [ 773.718904] env[61728]: _type = "Task" [ 773.718904] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.728679] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522ec186-a660-865e-6b6f-45e96c32aef2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.730585] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.710s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.731695] env[61728]: DEBUG nova.compute.manager [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 773.733887] env[61728]: DEBUG oslo_concurrency.lockutils [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 45.147s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.734278] env[61728]: DEBUG nova.objects.instance [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lazy-loading 'resources' on Instance uuid c8f1f4ae-ab08-4a03-a3fd-94014509e7e7 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 773.993999] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 773.993999] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122092', 'volume_id': '11c73a47-a9c0-4b47-a431-15f14807f4f9', 'name': 'volume-11c73a47-a9c0-4b47-a431-15f14807f4f9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c84bfcd0-b145-4675-8b0a-5e8f94f65098', 'attached_at': '', 'detached_at': '', 'volume_id': '11c73a47-a9c0-4b47-a431-15f14807f4f9', 'serial': '11c73a47-a9c0-4b47-a431-15f14807f4f9'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 773.994976] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-376f1afe-8a0b-4da1-95aa-48b8e740721f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.013972] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a75a1c9-a608-4b1c-8351-8931299354fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.045914] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] volume-11c73a47-a9c0-4b47-a431-15f14807f4f9/volume-11c73a47-a9c0-4b47-a431-15f14807f4f9.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 774.046411] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98a88450-9624-4d23-8ac6-5252f63cf7f6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.074263] env[61728]: DEBUG nova.network.neutron [-] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.076866] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464142, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096794} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.078457] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 774.078895] env[61728]: DEBUG oslo_vmware.api [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 774.078895] env[61728]: value = "task-464143" [ 774.078895] env[61728]: _type = "Task" [ 774.078895] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.079710] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9652470f-8c53-423e-97c7-8c9bd35da888 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.111710] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 06e765e1-85d9-4f6c-8ab5-81735f08cb66/06e765e1-85d9-4f6c-8ab5-81735f08cb66.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 774.118234] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee353602-7bbd-476b-a107-d5409bab4240 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.139947] env[61728]: DEBUG nova.compute.manager [req-dd7b8c94-f777-48ec-9a26-5ceac9417120 req-d5d212f1-4410-435d-93aa-29019cf4a897 service nova] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Received event network-changed-6f900c64-beea-4e8a-84c5-9f13768f8420 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 774.139947] env[61728]: DEBUG nova.compute.manager [req-dd7b8c94-f777-48ec-9a26-5ceac9417120 req-d5d212f1-4410-435d-93aa-29019cf4a897 service nova] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Refreshing instance network info cache due to event network-changed-6f900c64-beea-4e8a-84c5-9f13768f8420. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 774.139947] env[61728]: DEBUG oslo_concurrency.lockutils [req-dd7b8c94-f777-48ec-9a26-5ceac9417120 req-d5d212f1-4410-435d-93aa-29019cf4a897 service nova] Acquiring lock "refresh_cache-a5c7e3ff-6829-4528-9279-e865a0eb8512" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.139947] env[61728]: DEBUG oslo_concurrency.lockutils [req-dd7b8c94-f777-48ec-9a26-5ceac9417120 req-d5d212f1-4410-435d-93aa-29019cf4a897 service nova] Acquired lock "refresh_cache-a5c7e3ff-6829-4528-9279-e865a0eb8512" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.139947] env[61728]: DEBUG nova.network.neutron [req-dd7b8c94-f777-48ec-9a26-5ceac9417120 req-d5d212f1-4410-435d-93aa-29019cf4a897 service nova] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Refreshing network info cache for port 6f900c64-beea-4e8a-84c5-9f13768f8420 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 774.140829] env[61728]: DEBUG oslo_vmware.api [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464143, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.141628] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "b63fa3df-317a-46d4-b8e9-74e9e287efde" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.141872] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "b63fa3df-317a-46d4-b8e9-74e9e287efde" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.142210] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "b63fa3df-317a-46d4-b8e9-74e9e287efde-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.142385] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "b63fa3df-317a-46d4-b8e9-74e9e287efde-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.142717] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "b63fa3df-317a-46d4-b8e9-74e9e287efde-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.145695] env[61728]: INFO nova.compute.manager [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Terminating instance [ 774.148328] env[61728]: DEBUG nova.compute.manager [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 774.148328] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 774.149879] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5ad181-bc36-4066-a5a5-f9212eb51799 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.157661] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 774.157661] env[61728]: value = "task-464144" [ 774.157661] env[61728]: _type = "Task" [ 774.157661] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.164299] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 774.164299] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dab37c9a-f959-4a94-b917-89151dcc2309 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.169469] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464144, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.176949] env[61728]: DEBUG oslo_vmware.api [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 774.176949] env[61728]: value = "task-464145" [ 774.176949] env[61728]: _type = "Task" [ 774.176949] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.190564] env[61728]: DEBUG oslo_vmware.api [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464145, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.231280] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522ec186-a660-865e-6b6f-45e96c32aef2, 'name': SearchDatastore_Task, 'duration_secs': 0.012319} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.231855] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.235018] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 774.235018] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.235018] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.235018] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 774.235018] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-838a850f-61a1-4d9b-aa70-9e78857c6adc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.237399] env[61728]: DEBUG nova.compute.utils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.242023] env[61728]: DEBUG nova.compute.manager [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 774.242202] env[61728]: DEBUG nova.network.neutron [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 774.246310] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 774.246310] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 774.247146] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3876d80e-a3c0-4f88-a105-c4ed7275d782 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.254029] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Waiting for the task: (returnval){ [ 774.254029] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52956981-1fd2-d5e7-545e-ffe47219c940" [ 774.254029] env[61728]: _type = "Task" [ 774.254029] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.264512] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52956981-1fd2-d5e7-545e-ffe47219c940, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.322215] env[61728]: DEBUG nova.policy [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e51efcf23db4d5496edbc3b7104b4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32bb35e1dfce40e48be08bb568d3f2b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 774.413533] env[61728]: DEBUG nova.network.neutron [-] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.578796] env[61728]: INFO nova.compute.manager [-] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Took 1.63 seconds to deallocate network for instance. [ 774.596485] env[61728]: DEBUG oslo_vmware.api [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464143, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.670600] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464144, 'name': ReconfigVM_Task, 'duration_secs': 0.391043} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.671157] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 06e765e1-85d9-4f6c-8ab5-81735f08cb66/06e765e1-85d9-4f6c-8ab5-81735f08cb66.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 774.672059] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7075fcf5-f94e-4e52-8139-58b38dc380ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.690143] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 774.690143] env[61728]: value = "task-464146" [ 774.690143] env[61728]: _type = "Task" [ 774.690143] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.696834] env[61728]: DEBUG oslo_vmware.api [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464145, 'name': PowerOffVM_Task, 'duration_secs': 0.221525} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.699966] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 774.700175] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 774.701037] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e83c0c1-44f4-4049-bb07-3ea5371c7248 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.707552] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464146, 'name': Rename_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.715918] env[61728]: DEBUG nova.network.neutron [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Successfully created port: 0e719ade-f327-4ddb-8db5-8beda823ed1b {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 774.744356] env[61728]: DEBUG nova.compute.manager [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 774.768440] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52956981-1fd2-d5e7-545e-ffe47219c940, 'name': SearchDatastore_Task, 'duration_secs': 0.018203} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.775765] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f0df23b-8d5b-45f9-af11-ebe9d03b9c8c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.780946] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 774.781278] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 774.781563] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Deleting the datastore file [datastore1] b63fa3df-317a-46d4-b8e9-74e9e287efde {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 774.781918] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-292faebf-6498-44a2-82e2-0ac29fbd8192 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.788361] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Waiting for the task: (returnval){ [ 774.788361] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527dee87-c939-8184-bcdc-c78a17d0fb5c" [ 774.788361] env[61728]: _type = "Task" [ 774.788361] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.796598] env[61728]: DEBUG oslo_vmware.api [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 774.796598] env[61728]: value = "task-464148" [ 774.796598] env[61728]: _type = "Task" [ 774.796598] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.806143] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527dee87-c939-8184-bcdc-c78a17d0fb5c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.815399] env[61728]: DEBUG oslo_vmware.api [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464148, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.916121] env[61728]: INFO nova.compute.manager [-] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Took 1.77 seconds to deallocate network for instance. [ 774.932328] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa23f93-7ffa-4aee-b54e-92710ab29394 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.946043] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eccdeaaf-a781-475f-b35b-5059616d872b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.981360] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc914a6-2089-4feb-808e-7a8b4f409ebb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.990154] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d8e546-192a-40ae-b035-8e43163cd661 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.005154] env[61728]: DEBUG nova.compute.provider_tree [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.051030] env[61728]: DEBUG nova.network.neutron [req-dd7b8c94-f777-48ec-9a26-5ceac9417120 req-d5d212f1-4410-435d-93aa-29019cf4a897 service nova] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Updated VIF entry in instance network info cache for port 6f900c64-beea-4e8a-84c5-9f13768f8420. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 775.051436] env[61728]: DEBUG nova.network.neutron [req-dd7b8c94-f777-48ec-9a26-5ceac9417120 req-d5d212f1-4410-435d-93aa-29019cf4a897 service nova] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Updating instance_info_cache with network_info: [{"id": "6f900c64-beea-4e8a-84c5-9f13768f8420", "address": "fa:16:3e:3c:d1:e3", "network": {"id": "4d4f744b-c837-4ae8-9d82-7dce6a34b344", "bridge": "br-int", "label": "tempest-ServersTestJSON-1248642690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0776c511b94f2e88ded3b52f5a7b76", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f900c64-be", "ovs_interfaceid": "6f900c64-beea-4e8a-84c5-9f13768f8420", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.090487] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.096968] env[61728]: DEBUG oslo_vmware.api [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464143, 'name': ReconfigVM_Task, 'duration_secs': 0.853118} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.097283] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Reconfigured VM instance instance-00000033 to attach disk [datastore1] volume-11c73a47-a9c0-4b47-a431-15f14807f4f9/volume-11c73a47-a9c0-4b47-a431-15f14807f4f9.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 775.103433] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-860f7d73-d50b-4ab5-844c-531257506377 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.120057] env[61728]: DEBUG oslo_vmware.api [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 775.120057] env[61728]: value = "task-464149" [ 775.120057] env[61728]: _type = "Task" [ 775.120057] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.129073] env[61728]: DEBUG oslo_vmware.api [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464149, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.202023] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464146, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.306748] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527dee87-c939-8184-bcdc-c78a17d0fb5c, 'name': SearchDatastore_Task, 'duration_secs': 0.023207} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.307693] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.308151] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a5c7e3ff-6829-4528-9279-e865a0eb8512/a5c7e3ff-6829-4528-9279-e865a0eb8512.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 775.308768] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-58bf9e5e-d196-48b7-8409-639223b86785 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.317681] env[61728]: DEBUG oslo_vmware.api [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464148, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.323051} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.318550] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 775.319456] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 775.319770] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 775.320065] env[61728]: INFO nova.compute.manager [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Took 1.17 seconds to destroy the instance on the hypervisor. [ 775.320421] env[61728]: DEBUG oslo.service.loopingcall [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.320758] env[61728]: DEBUG nova.compute.manager [-] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 775.320982] env[61728]: DEBUG nova.network.neutron [-] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 775.324471] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Waiting for the task: (returnval){ [ 775.324471] env[61728]: value = "task-464150" [ 775.324471] env[61728]: _type = "Task" [ 775.324471] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.336613] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464150, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.476268] env[61728]: INFO nova.compute.manager [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Took 0.56 seconds to detach 1 volumes for instance. [ 775.479241] env[61728]: DEBUG nova.compute.manager [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Deleting volume: 2e85ae80-363c-4fa7-add2-31dd95e23cfd {{(pid=61728) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 775.510691] env[61728]: DEBUG nova.scheduler.client.report [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 775.553732] env[61728]: DEBUG oslo_concurrency.lockutils [req-dd7b8c94-f777-48ec-9a26-5ceac9417120 req-d5d212f1-4410-435d-93aa-29019cf4a897 service nova] Releasing lock "refresh_cache-a5c7e3ff-6829-4528-9279-e865a0eb8512" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.554035] env[61728]: DEBUG nova.compute.manager [req-dd7b8c94-f777-48ec-9a26-5ceac9417120 req-d5d212f1-4410-435d-93aa-29019cf4a897 service nova] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Received event network-vif-deleted-9666e319-339d-4b02-93fa-ec1aaf79fb7b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 775.632064] env[61728]: DEBUG oslo_vmware.api [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464149, 'name': ReconfigVM_Task, 'duration_secs': 0.162761} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.632451] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122092', 'volume_id': '11c73a47-a9c0-4b47-a431-15f14807f4f9', 'name': 'volume-11c73a47-a9c0-4b47-a431-15f14807f4f9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c84bfcd0-b145-4675-8b0a-5e8f94f65098', 'attached_at': '', 'detached_at': '', 'volume_id': '11c73a47-a9c0-4b47-a431-15f14807f4f9', 'serial': '11c73a47-a9c0-4b47-a431-15f14807f4f9'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 775.706086] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464146, 'name': Rename_Task, 'duration_secs': 0.958418} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.707114] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 775.707114] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-50c68535-2409-4d00-9027-e0740d5bf271 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.717209] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 775.717209] env[61728]: value = "task-464152" [ 775.717209] env[61728]: _type = "Task" [ 775.717209] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.730026] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464152, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.756190] env[61728]: DEBUG nova.compute.manager [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 775.779301] env[61728]: DEBUG nova.virt.hardware [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 775.779610] env[61728]: DEBUG nova.virt.hardware [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 775.779775] env[61728]: DEBUG nova.virt.hardware [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.779968] env[61728]: DEBUG nova.virt.hardware [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 775.780140] env[61728]: DEBUG nova.virt.hardware [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.780300] env[61728]: DEBUG nova.virt.hardware [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 775.780523] env[61728]: DEBUG nova.virt.hardware [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 775.780688] env[61728]: DEBUG nova.virt.hardware [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 775.780863] env[61728]: DEBUG nova.virt.hardware [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 775.781039] env[61728]: DEBUG nova.virt.hardware [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 775.781225] env[61728]: DEBUG nova.virt.hardware [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 775.782516] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad94603b-bc48-424b-b1d9-3a96f097b572 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.791842] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd7c4379-cec8-4367-a20e-ccb1b7fd2abf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.839684] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464150, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.018033] env[61728]: DEBUG oslo_concurrency.lockutils [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.282s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.018962] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 42.170s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.021347] env[61728]: INFO nova.compute.claims [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.035540] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.045222] env[61728]: INFO nova.scheduler.client.report [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Deleted allocations for instance c8f1f4ae-ab08-4a03-a3fd-94014509e7e7 [ 776.093292] env[61728]: DEBUG nova.network.neutron [-] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.218041] env[61728]: DEBUG nova.compute.manager [req-1503f64d-d77f-481a-89de-686ef1d53582 req-93d55a1a-e072-4bd2-bd4e-a926687b4e2d service nova] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Received event network-vif-deleted-b797bbf1-77c6-48d3-b63c-f4caf481f773 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 776.218256] env[61728]: DEBUG nova.compute.manager [req-1503f64d-d77f-481a-89de-686ef1d53582 req-93d55a1a-e072-4bd2-bd4e-a926687b4e2d service nova] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Received event network-vif-deleted-3f4753c4-b820-4467-8f8a-6084733c6e28 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 777.003727] env[61728]: DEBUG nova.network.neutron [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Successfully updated port: 0e719ade-f327-4ddb-8db5-8beda823ed1b {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 777.005568] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.005882] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.006130] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.006454] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.006508] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.013048] env[61728]: INFO nova.compute.manager [-] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Took 1.69 seconds to deallocate network for instance. [ 777.013048] env[61728]: DEBUG nova.objects.instance [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lazy-loading 'flavor' on Instance uuid c84bfcd0-b145-4675-8b0a-5e8f94f65098 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 777.014659] env[61728]: INFO nova.compute.manager [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Terminating instance [ 777.017899] env[61728]: DEBUG oslo_concurrency.lockutils [None req-18e22e3d-a865-4abe-8ee1-726ebc448a46 tempest-ServersAdminNegativeTestJSON-1060434196 tempest-ServersAdminNegativeTestJSON-1060434196-project-member] Lock "c8f1f4ae-ab08-4a03-a3fd-94014509e7e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 51.355s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.024962] env[61728]: DEBUG nova.compute.manager [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 777.025104] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 777.026233] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8cf8e1-4d7a-4d2d-aa43-eb33fd820d08 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.033051] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464152, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.033279] env[61728]: WARNING oslo_vmware.common.loopingcall [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] task run outlasted interval by 0.31466799999999995 sec [ 777.043749] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464150, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.563759} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.043887] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 777.044588] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a5c7e3ff-6829-4528-9279-e865a0eb8512/a5c7e3ff-6829-4528-9279-e865a0eb8512.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 777.044815] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 777.045067] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac0326cd-c907-4651-8f3a-00305992cc50 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.046881] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bf6acc68-e1e3-4da2-8fef-8005bf93978a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.052798] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464152, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.064048] env[61728]: DEBUG oslo_vmware.api [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 777.064048] env[61728]: value = "task-464153" [ 777.064048] env[61728]: _type = "Task" [ 777.064048] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.064385] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Waiting for the task: (returnval){ [ 777.064385] env[61728]: value = "task-464154" [ 777.064385] env[61728]: _type = "Task" [ 777.064385] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.085481] env[61728]: DEBUG oslo_vmware.api [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-464153, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.085481] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464154, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.518887] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.519279] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.519279] env[61728]: DEBUG nova.network.neutron [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 777.525216] env[61728]: DEBUG oslo_concurrency.lockutils [None req-591da1b2-81de-4c28-bd87-adb71e76d181 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.673s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.531995] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.548038] env[61728]: DEBUG oslo_vmware.api [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464152, 'name': PowerOnVM_Task, 'duration_secs': 1.681921} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.548289] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 777.548558] env[61728]: INFO nova.compute.manager [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Took 9.35 seconds to spawn the instance on the hypervisor. [ 777.548759] env[61728]: DEBUG nova.compute.manager [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 777.552370] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96486e4c-16e9-44ff-a3c0-7ea69562d2e3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.579885] env[61728]: DEBUG oslo_vmware.api [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-464153, 'name': PowerOffVM_Task, 'duration_secs': 0.21335} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.583879] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 777.584282] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 777.584363] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464154, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.116824} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.586943] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1ee53c9-0d26-45dd-8dc7-72d20707f360 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.588635] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 777.589532] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1822b7-4456-4429-a0cc-d2d00671a79a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.617143] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] a5c7e3ff-6829-4528-9279-e865a0eb8512/a5c7e3ff-6829-4528-9279-e865a0eb8512.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 777.621345] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02e0c44b-43e1-43a5-9c50-0f58092bc8b7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.646148] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Waiting for the task: (returnval){ [ 777.646148] env[61728]: value = "task-464156" [ 777.646148] env[61728]: _type = "Task" [ 777.646148] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.664214] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464156, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.667126] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 777.667603] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 777.667821] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Deleting the datastore file [datastore1] c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 777.668148] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e39e7d42-5f3d-4a4f-8ae3-320261831da2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.678977] env[61728]: DEBUG oslo_vmware.api [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for the task: (returnval){ [ 777.678977] env[61728]: value = "task-464157" [ 777.678977] env[61728]: _type = "Task" [ 777.678977] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.685631] env[61728]: DEBUG oslo_vmware.api [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-464157, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.873993] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.874395] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.061395] env[61728]: DEBUG nova.network.neutron [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.073152] env[61728]: INFO nova.compute.manager [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Took 58.48 seconds to build instance. [ 778.159341] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464156, 'name': ReconfigVM_Task, 'duration_secs': 0.328097} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.163092] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Reconfigured VM instance instance-0000003c to attach disk [datastore1] a5c7e3ff-6829-4528-9279-e865a0eb8512/a5c7e3ff-6829-4528-9279-e865a0eb8512.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 778.164777] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-20b1d6e3-880d-4cb9-8c61-4fe890ef2e14 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.175336] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Waiting for the task: (returnval){ [ 778.175336] env[61728]: value = "task-464158" [ 778.175336] env[61728]: _type = "Task" [ 778.175336] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.181272] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab2cc25-799b-4954-943b-651c5c7c6a25 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.194803] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464158, 'name': Rename_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.200096] env[61728]: DEBUG oslo_vmware.api [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Task: {'id': task-464157, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.236505} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.200456] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 778.200682] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 778.200871] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 778.201082] env[61728]: INFO nova.compute.manager [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Took 1.18 seconds to destroy the instance on the hypervisor. [ 778.201368] env[61728]: DEBUG oslo.service.loopingcall [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.202657] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90396161-4d2b-4e43-ad18-dd50dab2b170 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.206049] env[61728]: DEBUG nova.compute.manager [-] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.206199] env[61728]: DEBUG nova.network.neutron [-] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 778.238887] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d79b4f6a-91bb-424c-a127-e63264f50b27 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.248155] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82dc80f7-1c4e-4943-95dd-1308da3f1f5c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.263163] env[61728]: DEBUG nova.compute.provider_tree [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.378175] env[61728]: INFO nova.compute.manager [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Detaching volume 11c73a47-a9c0-4b47-a431-15f14807f4f9 [ 778.417469] env[61728]: INFO nova.virt.block_device [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Attempting to driver detach volume 11c73a47-a9c0-4b47-a431-15f14807f4f9 from mountpoint /dev/sdb [ 778.417797] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 778.418059] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122092', 'volume_id': '11c73a47-a9c0-4b47-a431-15f14807f4f9', 'name': 'volume-11c73a47-a9c0-4b47-a431-15f14807f4f9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c84bfcd0-b145-4675-8b0a-5e8f94f65098', 'attached_at': '', 'detached_at': '', 'volume_id': '11c73a47-a9c0-4b47-a431-15f14807f4f9', 'serial': '11c73a47-a9c0-4b47-a431-15f14807f4f9'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 778.419154] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64d536a-c8c4-4325-902b-87efbf5863aa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.449804] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bea571b-1ab1-43bd-8e3c-607c9951c975 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.460788] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af74cd1c-0261-47d9-a116-786d63e51baf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.491782] env[61728]: DEBUG nova.network.neutron [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Updating instance_info_cache with network_info: [{"id": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "address": "fa:16:3e:44:e0:e6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e719ade-f3", "ovs_interfaceid": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.496205] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5a1e76-41f1-4fc6-9e74-b150b77af372 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.514279] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] The volume has not been displaced from its original location: [datastore1] volume-11c73a47-a9c0-4b47-a431-15f14807f4f9/volume-11c73a47-a9c0-4b47-a431-15f14807f4f9.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 778.520137] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Reconfiguring VM instance instance-00000033 to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 778.522108] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f362884a-adeb-4438-b368-379325e395d9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.536795] env[61728]: DEBUG nova.compute.manager [req-33f5af58-b25f-48db-b0e0-cf7b5c9c3b50 req-8b477e4b-5d6b-460b-afe4-be53e3fac40a service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Received event network-vif-plugged-0e719ade-f327-4ddb-8db5-8beda823ed1b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 778.536795] env[61728]: DEBUG oslo_concurrency.lockutils [req-33f5af58-b25f-48db-b0e0-cf7b5c9c3b50 req-8b477e4b-5d6b-460b-afe4-be53e3fac40a service nova] Acquiring lock "f28348d2-c062-497a-b374-521df51054ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.536795] env[61728]: DEBUG oslo_concurrency.lockutils [req-33f5af58-b25f-48db-b0e0-cf7b5c9c3b50 req-8b477e4b-5d6b-460b-afe4-be53e3fac40a service nova] Lock "f28348d2-c062-497a-b374-521df51054ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.536795] env[61728]: DEBUG oslo_concurrency.lockutils [req-33f5af58-b25f-48db-b0e0-cf7b5c9c3b50 req-8b477e4b-5d6b-460b-afe4-be53e3fac40a service nova] Lock "f28348d2-c062-497a-b374-521df51054ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.536795] env[61728]: DEBUG nova.compute.manager [req-33f5af58-b25f-48db-b0e0-cf7b5c9c3b50 req-8b477e4b-5d6b-460b-afe4-be53e3fac40a service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] No waiting events found dispatching network-vif-plugged-0e719ade-f327-4ddb-8db5-8beda823ed1b {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 778.536959] env[61728]: WARNING nova.compute.manager [req-33f5af58-b25f-48db-b0e0-cf7b5c9c3b50 req-8b477e4b-5d6b-460b-afe4-be53e3fac40a service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Received unexpected event network-vif-plugged-0e719ade-f327-4ddb-8db5-8beda823ed1b for instance with vm_state building and task_state spawning. [ 778.537166] env[61728]: DEBUG nova.compute.manager [req-33f5af58-b25f-48db-b0e0-cf7b5c9c3b50 req-8b477e4b-5d6b-460b-afe4-be53e3fac40a service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Received event network-changed-0e719ade-f327-4ddb-8db5-8beda823ed1b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 778.537235] env[61728]: DEBUG nova.compute.manager [req-33f5af58-b25f-48db-b0e0-cf7b5c9c3b50 req-8b477e4b-5d6b-460b-afe4-be53e3fac40a service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Refreshing instance network info cache due to event network-changed-0e719ade-f327-4ddb-8db5-8beda823ed1b. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 778.537387] env[61728]: DEBUG oslo_concurrency.lockutils [req-33f5af58-b25f-48db-b0e0-cf7b5c9c3b50 req-8b477e4b-5d6b-460b-afe4-be53e3fac40a service nova] Acquiring lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.546917] env[61728]: DEBUG oslo_vmware.api [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 778.546917] env[61728]: value = "task-464159" [ 778.546917] env[61728]: _type = "Task" [ 778.546917] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.560669] env[61728]: DEBUG oslo_vmware.api [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464159, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.578370] env[61728]: DEBUG oslo_concurrency.lockutils [None req-646929bf-353a-4bfd-a1f2-9c89d188e136 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.352s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.689977] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464158, 'name': Rename_Task, 'duration_secs': 0.426363} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.691120] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 778.691120] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-534bbd44-d76a-4be8-84ed-92df926b3ea1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.701771] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Waiting for the task: (returnval){ [ 778.701771] env[61728]: value = "task-464160" [ 778.701771] env[61728]: _type = "Task" [ 778.701771] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.714275] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464160, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.766848] env[61728]: DEBUG nova.scheduler.client.report [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.000702] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.001048] env[61728]: DEBUG nova.compute.manager [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Instance network_info: |[{"id": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "address": "fa:16:3e:44:e0:e6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e719ade-f3", "ovs_interfaceid": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 779.001441] env[61728]: DEBUG oslo_concurrency.lockutils [req-33f5af58-b25f-48db-b0e0-cf7b5c9c3b50 req-8b477e4b-5d6b-460b-afe4-be53e3fac40a service nova] Acquired lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.001794] env[61728]: DEBUG nova.network.neutron [req-33f5af58-b25f-48db-b0e0-cf7b5c9c3b50 req-8b477e4b-5d6b-460b-afe4-be53e3fac40a service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Refreshing network info cache for port 0e719ade-f327-4ddb-8db5-8beda823ed1b {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 779.003233] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:e0:e6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0e719ade-f327-4ddb-8db5-8beda823ed1b', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 779.013388] env[61728]: DEBUG oslo.service.loopingcall [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.014654] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f28348d2-c062-497a-b374-521df51054ee] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 779.015811] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a235fe0a-17c9-4dc1-895d-8c5c8c12ec35 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.040756] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 779.040756] env[61728]: value = "task-464161" [ 779.040756] env[61728]: _type = "Task" [ 779.040756] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.054879] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464161, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.060492] env[61728]: DEBUG oslo_vmware.api [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464159, 'name': ReconfigVM_Task, 'duration_secs': 0.307915} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.061412] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Reconfigured VM instance instance-00000033 to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 779.066157] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a8f4888-ab7e-4c95-83f4-7be03156be3b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.102191] env[61728]: DEBUG oslo_vmware.api [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 779.102191] env[61728]: value = "task-464162" [ 779.102191] env[61728]: _type = "Task" [ 779.102191] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.122083] env[61728]: DEBUG oslo_vmware.api [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464162, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.182919] env[61728]: DEBUG nova.network.neutron [-] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.214479] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464160, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.272979] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.254s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.273819] env[61728]: DEBUG nova.compute.manager [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 779.278036] env[61728]: DEBUG oslo_concurrency.lockutils [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.398s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.278398] env[61728]: DEBUG nova.objects.instance [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lazy-loading 'resources' on Instance uuid c14034b5-330b-4fb1-af31-d530e3231992 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 779.550932] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464161, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.617552] env[61728]: DEBUG oslo_vmware.api [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464162, 'name': ReconfigVM_Task, 'duration_secs': 0.237115} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.617805] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122092', 'volume_id': '11c73a47-a9c0-4b47-a431-15f14807f4f9', 'name': 'volume-11c73a47-a9c0-4b47-a431-15f14807f4f9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c84bfcd0-b145-4675-8b0a-5e8f94f65098', 'attached_at': '', 'detached_at': '', 'volume_id': '11c73a47-a9c0-4b47-a431-15f14807f4f9', 'serial': '11c73a47-a9c0-4b47-a431-15f14807f4f9'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 779.685667] env[61728]: INFO nova.compute.manager [-] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Took 1.48 seconds to deallocate network for instance. [ 779.714064] env[61728]: DEBUG oslo_vmware.api [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464160, 'name': PowerOnVM_Task, 'duration_secs': 0.574812} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.714325] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 779.714559] env[61728]: INFO nova.compute.manager [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Took 8.91 seconds to spawn the instance on the hypervisor. [ 779.714710] env[61728]: DEBUG nova.compute.manager [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 779.715546] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a6a126-9c33-4800-bd44-47b488b4f76d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.782515] env[61728]: DEBUG nova.compute.utils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 779.785962] env[61728]: DEBUG nova.compute.manager [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 779.786176] env[61728]: DEBUG nova.network.neutron [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 779.889417] env[61728]: DEBUG nova.policy [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f8738985bd14fc3a811876fa6041674', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '037e1941b7964a2b94d5b0f5c280956e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 779.929354] env[61728]: DEBUG nova.network.neutron [req-33f5af58-b25f-48db-b0e0-cf7b5c9c3b50 req-8b477e4b-5d6b-460b-afe4-be53e3fac40a service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Updated VIF entry in instance network info cache for port 0e719ade-f327-4ddb-8db5-8beda823ed1b. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 779.929946] env[61728]: DEBUG nova.network.neutron [req-33f5af58-b25f-48db-b0e0-cf7b5c9c3b50 req-8b477e4b-5d6b-460b-afe4-be53e3fac40a service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Updating instance_info_cache with network_info: [{"id": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "address": "fa:16:3e:44:e0:e6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e719ade-f3", "ovs_interfaceid": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.051809] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464161, 'name': CreateVM_Task, 'duration_secs': 0.589116} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.051809] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f28348d2-c062-497a-b374-521df51054ee] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 780.052346] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.052515] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.052910] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 780.056858] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-029df744-59e6-40d4-a742-7b6e349756ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.062405] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 780.062405] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fbf59b-ea32-5155-9bca-6401d682235c" [ 780.062405] env[61728]: _type = "Task" [ 780.062405] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.076936] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fbf59b-ea32-5155-9bca-6401d682235c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.196950] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.198524] env[61728]: DEBUG nova.objects.instance [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lazy-loading 'flavor' on Instance uuid c84bfcd0-b145-4675-8b0a-5e8f94f65098 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 780.238214] env[61728]: INFO nova.compute.manager [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Took 56.60 seconds to build instance. [ 780.284869] env[61728]: DEBUG nova.compute.manager [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 780.434125] env[61728]: DEBUG oslo_concurrency.lockutils [req-33f5af58-b25f-48db-b0e0-cf7b5c9c3b50 req-8b477e4b-5d6b-460b-afe4-be53e3fac40a service nova] Releasing lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.464823] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3464eb08-4e0b-4ce6-a2a8-c9ad87cc33fe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.476087] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf20398-6e54-478c-af1b-37681374647d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.480128] env[61728]: DEBUG nova.network.neutron [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Successfully created port: c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.519271] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e016a7-5935-40ab-b73c-4d73b5d41a9a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.529406] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8880854e-5137-4152-a8ed-30686ced112f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.545745] env[61728]: DEBUG nova.compute.provider_tree [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.576546] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fbf59b-ea32-5155-9bca-6401d682235c, 'name': SearchDatastore_Task, 'duration_secs': 0.014784} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.576915] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.577189] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 780.577565] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.577624] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.578915] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 780.578915] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1bff8b6d-c415-4858-b9e8-3b1a5fb9acd4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.589736] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 780.589736] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 780.589736] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e82b6a8d-5fbf-451f-85fa-c46f6045e49b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.597439] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 780.597439] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52837aa1-75d8-5782-71f2-dc27a4815bb9" [ 780.597439] env[61728]: _type = "Task" [ 780.597439] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.607588] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52837aa1-75d8-5782-71f2-dc27a4815bb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.683827] env[61728]: DEBUG nova.compute.manager [req-4f18f89c-4a9e-4bf4-a2d5-0e89740c6f54 req-54e2e711-29c3-4504-b8a2-e271c76a1beb service nova] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Received event network-vif-deleted-ccffc76c-1df1-4ba7-824a-5c175febeca3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 780.684247] env[61728]: DEBUG nova.compute.manager [req-4f18f89c-4a9e-4bf4-a2d5-0e89740c6f54 req-54e2e711-29c3-4504-b8a2-e271c76a1beb service nova] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Received event network-changed-33a9ae57-3181-42d6-9a7f-a66bf8103138 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 780.684538] env[61728]: DEBUG nova.compute.manager [req-4f18f89c-4a9e-4bf4-a2d5-0e89740c6f54 req-54e2e711-29c3-4504-b8a2-e271c76a1beb service nova] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Refreshing instance network info cache due to event network-changed-33a9ae57-3181-42d6-9a7f-a66bf8103138. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 780.684876] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f18f89c-4a9e-4bf4-a2d5-0e89740c6f54 req-54e2e711-29c3-4504-b8a2-e271c76a1beb service nova] Acquiring lock "refresh_cache-06e765e1-85d9-4f6c-8ab5-81735f08cb66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.685158] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f18f89c-4a9e-4bf4-a2d5-0e89740c6f54 req-54e2e711-29c3-4504-b8a2-e271c76a1beb service nova] Acquired lock "refresh_cache-06e765e1-85d9-4f6c-8ab5-81735f08cb66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.685425] env[61728]: DEBUG nova.network.neutron [req-4f18f89c-4a9e-4bf4-a2d5-0e89740c6f54 req-54e2e711-29c3-4504-b8a2-e271c76a1beb service nova] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Refreshing network info cache for port 33a9ae57-3181-42d6-9a7f-a66bf8103138 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 780.743110] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eedd215-6666-46c5-9e3f-f55fae645cb7 tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Lock "a5c7e3ff-6829-4528-9279-e865a0eb8512" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.757s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.993550] env[61728]: DEBUG nova.network.neutron [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Successfully created port: 60b8853e-0428-415e-99cb-0a8541ea7a64 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 781.052035] env[61728]: DEBUG nova.scheduler.client.report [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.115024] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52837aa1-75d8-5782-71f2-dc27a4815bb9, 'name': SearchDatastore_Task, 'duration_secs': 0.015978} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.115024] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edaed4cd-eb8f-4c00-94a9-1e915afdb7d8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.120493] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 781.120493] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525f13a4-4229-3fb7-1ed4-99f63bc03001" [ 781.120493] env[61728]: _type = "Task" [ 781.120493] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.133637] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525f13a4-4229-3fb7-1ed4-99f63bc03001, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.209176] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7058e69d-5652-48f3-8418-7e2b7f6ee8c6 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.334s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.297144] env[61728]: DEBUG nova.compute.manager [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 781.330164] env[61728]: DEBUG nova.virt.hardware [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.330569] env[61728]: DEBUG nova.virt.hardware [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.330793] env[61728]: DEBUG nova.virt.hardware [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.331056] env[61728]: DEBUG nova.virt.hardware [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.331299] env[61728]: DEBUG nova.virt.hardware [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.331528] env[61728]: DEBUG nova.virt.hardware [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.331812] env[61728]: DEBUG nova.virt.hardware [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.332087] env[61728]: DEBUG nova.virt.hardware [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.332318] env[61728]: DEBUG nova.virt.hardware [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.332554] env[61728]: DEBUG nova.virt.hardware [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.332859] env[61728]: DEBUG nova.virt.hardware [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.333968] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78267697-98ee-4b3b-b01c-527157cc54c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.349676] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ebc6cf-277d-4ff3-b1d3-d68364f01661 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.487526] env[61728]: DEBUG nova.network.neutron [req-4f18f89c-4a9e-4bf4-a2d5-0e89740c6f54 req-54e2e711-29c3-4504-b8a2-e271c76a1beb service nova] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Updated VIF entry in instance network info cache for port 33a9ae57-3181-42d6-9a7f-a66bf8103138. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 781.488060] env[61728]: DEBUG nova.network.neutron [req-4f18f89c-4a9e-4bf4-a2d5-0e89740c6f54 req-54e2e711-29c3-4504-b8a2-e271c76a1beb service nova] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Updating instance_info_cache with network_info: [{"id": "33a9ae57-3181-42d6-9a7f-a66bf8103138", "address": "fa:16:3e:9c:55:7f", "network": {"id": "22ad4b68-2895-4651-803f-44b8031a6292", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-489227259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "276942c8cf8a42729a541de096f69a0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33a9ae57-31", "ovs_interfaceid": "33a9ae57-3181-42d6-9a7f-a66bf8103138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.561311] env[61728]: DEBUG oslo_concurrency.lockutils [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.280s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.561311] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 40.947s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.584087] env[61728]: INFO nova.scheduler.client.report [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleted allocations for instance c14034b5-330b-4fb1-af31-d530e3231992 [ 781.635869] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525f13a4-4229-3fb7-1ed4-99f63bc03001, 'name': SearchDatastore_Task, 'duration_secs': 0.01681} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.638020] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.638020] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] f28348d2-c062-497a-b374-521df51054ee/f28348d2-c062-497a-b374-521df51054ee.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 781.638020] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49704d2f-373e-40b1-af70-026be00a84b4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.647672] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 781.647672] env[61728]: value = "task-464163" [ 781.647672] env[61728]: _type = "Task" [ 781.647672] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.659094] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464163, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.689678] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d76f18-e78c-3017-7ef0-ca2c35a87c0e/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 781.690230] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c68b23f-5638-4395-b105-19f6282a2760 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.698153] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d76f18-e78c-3017-7ef0-ca2c35a87c0e/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 781.698328] env[61728]: ERROR oslo_vmware.rw_handles [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d76f18-e78c-3017-7ef0-ca2c35a87c0e/disk-0.vmdk due to incomplete transfer. [ 781.702905] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4d890ae6-fdb7-4525-a542-2d409f3ebdb9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.711320] env[61728]: DEBUG oslo_vmware.rw_handles [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d76f18-e78c-3017-7ef0-ca2c35a87c0e/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 781.711560] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Uploaded image 05e97342-0afd-4bd8-b341-7f3386bdf448 to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 781.716551] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 781.717671] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9bdb42be-2722-4bfe-973e-43e0b1a4e23e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.726022] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 781.726022] env[61728]: value = "task-464164" [ 781.726022] env[61728]: _type = "Task" [ 781.726022] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.735220] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464164, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.994826] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f18f89c-4a9e-4bf4-a2d5-0e89740c6f54 req-54e2e711-29c3-4504-b8a2-e271c76a1beb service nova] Releasing lock "refresh_cache-06e765e1-85d9-4f6c-8ab5-81735f08cb66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.100270] env[61728]: DEBUG oslo_concurrency.lockutils [None req-57d66d3e-028a-4ee7-bd8a-95bc816dd11d tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "c14034b5-330b-4fb1-af31-d530e3231992" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 44.744s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.159672] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464163, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.238113] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464164, 'name': Destroy_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.258140] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b08b2149-8fad-4eeb-a4ac-a6dc025215f1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.266985] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ddddc4-22ae-4b66-ba66-cea6a10446f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.299637] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ab510f-589a-4b97-ba44-965f9fd999a3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.308145] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62247a95-8079-4acb-b766-8d37f88b0bec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.322774] env[61728]: DEBUG nova.compute.provider_tree [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.660174] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464163, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516271} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.660409] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] f28348d2-c062-497a-b374-521df51054ee/f28348d2-c062-497a-b374-521df51054ee.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 782.660628] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 782.661085] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a45e794c-ae8f-42ca-9519-071ab58f0655 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.671319] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 782.671319] env[61728]: value = "task-464165" [ 782.671319] env[61728]: _type = "Task" [ 782.671319] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.672636] env[61728]: DEBUG nova.network.neutron [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Successfully updated port: c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 782.685271] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464165, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.739326] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464164, 'name': Destroy_Task, 'duration_secs': 0.632459} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.739793] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Destroyed the VM [ 782.740542] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 782.740966] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b357c4c0-083a-4122-bde7-6fa921c897f9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.749714] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 782.749714] env[61728]: value = "task-464166" [ 782.749714] env[61728]: _type = "Task" [ 782.749714] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.758937] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464166, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.825635] env[61728]: DEBUG nova.scheduler.client.report [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.031023] env[61728]: DEBUG nova.compute.manager [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Received event network-changed-6f900c64-beea-4e8a-84c5-9f13768f8420 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 783.031427] env[61728]: DEBUG nova.compute.manager [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Refreshing instance network info cache due to event network-changed-6f900c64-beea-4e8a-84c5-9f13768f8420. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 783.031948] env[61728]: DEBUG oslo_concurrency.lockutils [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] Acquiring lock "refresh_cache-a5c7e3ff-6829-4528-9279-e865a0eb8512" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.032240] env[61728]: DEBUG oslo_concurrency.lockutils [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] Acquired lock "refresh_cache-a5c7e3ff-6829-4528-9279-e865a0eb8512" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.032473] env[61728]: DEBUG nova.network.neutron [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Refreshing network info cache for port 6f900c64-beea-4e8a-84c5-9f13768f8420 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 783.187316] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464165, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074252} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.190101] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 783.194535] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8660943-17d8-4e75-8c32-3ca331d0044d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.198857] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "b444a6cf-3c98-4308-afc1-6e760d30082d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.199109] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "b444a6cf-3c98-4308-afc1-6e760d30082d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.224189] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] f28348d2-c062-497a-b374-521df51054ee/f28348d2-c062-497a-b374-521df51054ee.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 783.225272] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71a2cb5b-8b5c-4a22-a5e4-a755406723da {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.249881] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 783.249881] env[61728]: value = "task-464167" [ 783.249881] env[61728]: _type = "Task" [ 783.249881] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.265882] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464166, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.269544] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464167, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.330772] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.770s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.330960] env[61728]: INFO nova.compute.manager [None req-8355f37c-04d4-4761-b33b-f63db15fad7d tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Successfully reverted task state from rebuilding on failure for instance. [ 783.339137] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 41.201s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.339137] env[61728]: INFO nova.compute.claims [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.704728] env[61728]: DEBUG nova.compute.manager [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 783.762867] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464166, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.768568] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464167, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.770759] env[61728]: DEBUG nova.network.neutron [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Updated VIF entry in instance network info cache for port 6f900c64-beea-4e8a-84c5-9f13768f8420. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 783.771129] env[61728]: DEBUG nova.network.neutron [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Updating instance_info_cache with network_info: [{"id": "6f900c64-beea-4e8a-84c5-9f13768f8420", "address": "fa:16:3e:3c:d1:e3", "network": {"id": "4d4f744b-c837-4ae8-9d82-7dce6a34b344", "bridge": "br-int", "label": "tempest-ServersTestJSON-1248642690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.179", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e0776c511b94f2e88ded3b52f5a7b76", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f900c64-be", "ovs_interfaceid": "6f900c64-beea-4e8a-84c5-9f13768f8420", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.232040] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.262849] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464166, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.271019] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464167, 'name': ReconfigVM_Task, 'duration_secs': 0.782758} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.271019] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Reconfigured VM instance instance-0000003d to attach disk [datastore1] f28348d2-c062-497a-b374-521df51054ee/f28348d2-c062-497a-b374-521df51054ee.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.271709] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e8f19cd-cc19-4d08-a661-13b5736a1089 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.278036] env[61728]: DEBUG oslo_concurrency.lockutils [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] Releasing lock "refresh_cache-a5c7e3ff-6829-4528-9279-e865a0eb8512" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.278036] env[61728]: DEBUG nova.compute.manager [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Received event network-vif-plugged-c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 784.278036] env[61728]: DEBUG oslo_concurrency.lockutils [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] Acquiring lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.278036] env[61728]: DEBUG oslo_concurrency.lockutils [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] Lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.278036] env[61728]: DEBUG oslo_concurrency.lockutils [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] Lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.278036] env[61728]: DEBUG nova.compute.manager [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] No waiting events found dispatching network-vif-plugged-c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 784.278036] env[61728]: WARNING nova.compute.manager [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Received unexpected event network-vif-plugged-c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49 for instance with vm_state building and task_state spawning. [ 784.278036] env[61728]: DEBUG nova.compute.manager [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Received event network-changed-c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 784.278036] env[61728]: DEBUG nova.compute.manager [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Refreshing instance network info cache due to event network-changed-c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 784.278036] env[61728]: DEBUG oslo_concurrency.lockutils [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] Acquiring lock "refresh_cache-a716cf5a-4482-45cb-96ec-b8c38bc3e742" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.278036] env[61728]: DEBUG oslo_concurrency.lockutils [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] Acquired lock "refresh_cache-a716cf5a-4482-45cb-96ec-b8c38bc3e742" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.279147] env[61728]: DEBUG nova.network.neutron [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Refreshing network info cache for port c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 784.286136] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 784.286136] env[61728]: value = "task-464168" [ 784.286136] env[61728]: _type = "Task" [ 784.286136] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.297360] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464168, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.698865] env[61728]: DEBUG nova.network.neutron [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Successfully updated port: 60b8853e-0428-415e-99cb-0a8541ea7a64 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 784.761585] env[61728]: DEBUG oslo_vmware.api [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464166, 'name': RemoveSnapshot_Task, 'duration_secs': 1.525278} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.761898] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 784.762191] env[61728]: INFO nova.compute.manager [None req-b3addd1a-b64d-4482-9b0e-9139ff61ea10 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Took 16.61 seconds to snapshot the instance on the hypervisor. [ 784.803828] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464168, 'name': Rename_Task, 'duration_secs': 0.148151} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.803828] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 784.803828] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac3c8402-ca63-4c7b-a97f-b7d33aa4dbc1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.810965] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 784.810965] env[61728]: value = "task-464169" [ 784.810965] env[61728]: _type = "Task" [ 784.810965] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.812301] env[61728]: DEBUG nova.network.neutron [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 784.823713] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464169, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.896686] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5021152e-3c95-432c-be50-f460148dce27 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.904754] env[61728]: DEBUG nova.network.neutron [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.906688] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32067644-f9c4-4488-abe8-3946b4991a4a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.939428] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d2fc01c-3040-42e3-84f1-87652234f899 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.948882] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c300f0-f0d0-4647-a39c-49f859db6b01 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.964127] env[61728]: DEBUG nova.compute.provider_tree [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.202143] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "refresh_cache-a716cf5a-4482-45cb-96ec-b8c38bc3e742" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.293375] env[61728]: DEBUG nova.compute.manager [req-0ef67276-3a07-445d-ae1f-776b5b95c853 req-7b7bcbd4-8f62-443f-a8f5-484a446037c1 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Received event network-vif-plugged-60b8853e-0428-415e-99cb-0a8541ea7a64 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 785.293572] env[61728]: DEBUG oslo_concurrency.lockutils [req-0ef67276-3a07-445d-ae1f-776b5b95c853 req-7b7bcbd4-8f62-443f-a8f5-484a446037c1 service nova] Acquiring lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.293705] env[61728]: DEBUG oslo_concurrency.lockutils [req-0ef67276-3a07-445d-ae1f-776b5b95c853 req-7b7bcbd4-8f62-443f-a8f5-484a446037c1 service nova] Lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.293857] env[61728]: DEBUG oslo_concurrency.lockutils [req-0ef67276-3a07-445d-ae1f-776b5b95c853 req-7b7bcbd4-8f62-443f-a8f5-484a446037c1 service nova] Lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.294096] env[61728]: DEBUG nova.compute.manager [req-0ef67276-3a07-445d-ae1f-776b5b95c853 req-7b7bcbd4-8f62-443f-a8f5-484a446037c1 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] No waiting events found dispatching network-vif-plugged-60b8853e-0428-415e-99cb-0a8541ea7a64 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 785.294329] env[61728]: WARNING nova.compute.manager [req-0ef67276-3a07-445d-ae1f-776b5b95c853 req-7b7bcbd4-8f62-443f-a8f5-484a446037c1 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Received unexpected event network-vif-plugged-60b8853e-0428-415e-99cb-0a8541ea7a64 for instance with vm_state building and task_state spawning. [ 785.294502] env[61728]: DEBUG nova.compute.manager [req-0ef67276-3a07-445d-ae1f-776b5b95c853 req-7b7bcbd4-8f62-443f-a8f5-484a446037c1 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Received event network-changed-60b8853e-0428-415e-99cb-0a8541ea7a64 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 785.294661] env[61728]: DEBUG nova.compute.manager [req-0ef67276-3a07-445d-ae1f-776b5b95c853 req-7b7bcbd4-8f62-443f-a8f5-484a446037c1 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Refreshing instance network info cache due to event network-changed-60b8853e-0428-415e-99cb-0a8541ea7a64. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 785.294840] env[61728]: DEBUG oslo_concurrency.lockutils [req-0ef67276-3a07-445d-ae1f-776b5b95c853 req-7b7bcbd4-8f62-443f-a8f5-484a446037c1 service nova] Acquiring lock "refresh_cache-a716cf5a-4482-45cb-96ec-b8c38bc3e742" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.322314] env[61728]: DEBUG oslo_vmware.api [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464169, 'name': PowerOnVM_Task, 'duration_secs': 0.492906} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.322601] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 785.322807] env[61728]: INFO nova.compute.manager [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Took 9.57 seconds to spawn the instance on the hypervisor. [ 785.322992] env[61728]: DEBUG nova.compute.manager [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 785.323787] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5050f2df-2afa-469a-889a-64f90f1393df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.411733] env[61728]: DEBUG oslo_concurrency.lockutils [req-db902715-d5d2-4f15-a475-354562255ef9 req-651234b5-3427-48d9-9136-5bbd1d48e287 service nova] Releasing lock "refresh_cache-a716cf5a-4482-45cb-96ec-b8c38bc3e742" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.412415] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquired lock "refresh_cache-a716cf5a-4482-45cb-96ec-b8c38bc3e742" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.412415] env[61728]: DEBUG nova.network.neutron [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 785.467331] env[61728]: DEBUG nova.scheduler.client.report [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.843027] env[61728]: INFO nova.compute.manager [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Took 58.26 seconds to build instance. [ 785.959696] env[61728]: DEBUG nova.network.neutron [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 785.972225] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.636s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.972782] env[61728]: DEBUG nova.compute.manager [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.975886] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.146s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.976125] env[61728]: DEBUG nova.objects.instance [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Lazy-loading 'resources' on Instance uuid 0f75629b-bdec-4593-aceb-cd0478972ccc {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 786.345205] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c96becae-00ad-4e8e-aecf-af6647c46399 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "f28348d2-c062-497a-b374-521df51054ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.099s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.479239] env[61728]: DEBUG nova.compute.utils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.486451] env[61728]: DEBUG nova.compute.manager [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 786.486638] env[61728]: DEBUG nova.network.neutron [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 786.519928] env[61728]: DEBUG nova.network.neutron [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Updating instance_info_cache with network_info: [{"id": "c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49", "address": "fa:16:3e:ae:dc:02", "network": {"id": "4dbef1be-974e-49aa-af12-b623bc56a8fb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-266138646", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.248", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8a2ce4f-4d", "ovs_interfaceid": "c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "60b8853e-0428-415e-99cb-0a8541ea7a64", "address": "fa:16:3e:06:1b:f3", "network": {"id": "4a8d636f-6cc7-45af-acde-fa1c06e488b3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1818596159", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.164", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60b8853e-04", "ovs_interfaceid": "60b8853e-0428-415e-99cb-0a8541ea7a64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.561892] env[61728]: DEBUG nova.policy [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '82fb3a9473db4fcfa49c3305f0ac8ed7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f742e5f95a3f4ef8b407894687866958', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 786.837924] env[61728]: DEBUG nova.network.neutron [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Successfully created port: 920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.986746] env[61728]: DEBUG nova.compute.manager [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 787.027788] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Releasing lock "refresh_cache-a716cf5a-4482-45cb-96ec-b8c38bc3e742" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.028191] env[61728]: DEBUG nova.compute.manager [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Instance network_info: |[{"id": "c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49", "address": "fa:16:3e:ae:dc:02", "network": {"id": "4dbef1be-974e-49aa-af12-b623bc56a8fb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-266138646", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.248", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8a2ce4f-4d", "ovs_interfaceid": "c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "60b8853e-0428-415e-99cb-0a8541ea7a64", "address": "fa:16:3e:06:1b:f3", "network": {"id": "4a8d636f-6cc7-45af-acde-fa1c06e488b3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1818596159", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.164", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60b8853e-04", "ovs_interfaceid": "60b8853e-0428-415e-99cb-0a8541ea7a64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 787.028574] env[61728]: DEBUG oslo_concurrency.lockutils [req-0ef67276-3a07-445d-ae1f-776b5b95c853 req-7b7bcbd4-8f62-443f-a8f5-484a446037c1 service nova] Acquired lock "refresh_cache-a716cf5a-4482-45cb-96ec-b8c38bc3e742" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.028829] env[61728]: DEBUG nova.network.neutron [req-0ef67276-3a07-445d-ae1f-776b5b95c853 req-7b7bcbd4-8f62-443f-a8f5-484a446037c1 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Refreshing network info cache for port 60b8853e-0428-415e-99cb-0a8541ea7a64 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 787.031434] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:dc:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '42f08482-a1da-405d-9918-d733d9f5173c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:1b:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da0e5087-d65b-416f-90fe-beaa9c534ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '60b8853e-0428-415e-99cb-0a8541ea7a64', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 787.042869] env[61728]: DEBUG oslo.service.loopingcall [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.048687] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 787.049399] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ecc46238-8427-44e7-8e67-86cca0ebf3bf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.082581] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 787.082581] env[61728]: value = "task-464170" [ 787.082581] env[61728]: _type = "Task" [ 787.082581] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.094121] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464170, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.111575] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63e5c2d-fb75-473a-a555-7366cc448cdd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.120753] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5d6d0d-a6af-44bf-9b5f-87ad50bd4f84 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.160331] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22c8580-11e4-4fff-baa6-e5eb97564577 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.170888] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3cdf77-4d69-4ff8-81fe-95307ad9c205 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.193748] env[61728]: DEBUG nova.compute.provider_tree [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.593442] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464170, 'name': CreateVM_Task, 'duration_secs': 0.412182} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.595035] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 787.595035] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.595035] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.595035] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 787.595295] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be0d2064-07fc-4430-b0fa-2fe1a0a40fcc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.601487] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 787.601487] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525995b7-bb93-af31-91fa-ab6adc58cb82" [ 787.601487] env[61728]: _type = "Task" [ 787.601487] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.610922] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525995b7-bb93-af31-91fa-ab6adc58cb82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.699364] env[61728]: DEBUG nova.scheduler.client.report [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.834062] env[61728]: DEBUG nova.network.neutron [req-0ef67276-3a07-445d-ae1f-776b5b95c853 req-7b7bcbd4-8f62-443f-a8f5-484a446037c1 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Updated VIF entry in instance network info cache for port 60b8853e-0428-415e-99cb-0a8541ea7a64. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 787.834589] env[61728]: DEBUG nova.network.neutron [req-0ef67276-3a07-445d-ae1f-776b5b95c853 req-7b7bcbd4-8f62-443f-a8f5-484a446037c1 service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Updating instance_info_cache with network_info: [{"id": "c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49", "address": "fa:16:3e:ae:dc:02", "network": {"id": "4dbef1be-974e-49aa-af12-b623bc56a8fb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-266138646", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.248", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8a2ce4f-4d", "ovs_interfaceid": "c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "60b8853e-0428-415e-99cb-0a8541ea7a64", "address": "fa:16:3e:06:1b:f3", "network": {"id": "4a8d636f-6cc7-45af-acde-fa1c06e488b3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1818596159", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.164", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60b8853e-04", "ovs_interfaceid": "60b8853e-0428-415e-99cb-0a8541ea7a64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.999199] env[61728]: DEBUG nova.compute.manager [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 788.035846] env[61728]: DEBUG nova.virt.hardware [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.036146] env[61728]: DEBUG nova.virt.hardware [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.036315] env[61728]: DEBUG nova.virt.hardware [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.036503] env[61728]: DEBUG nova.virt.hardware [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.036650] env[61728]: DEBUG nova.virt.hardware [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.036798] env[61728]: DEBUG nova.virt.hardware [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.037019] env[61728]: DEBUG nova.virt.hardware [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.037265] env[61728]: DEBUG nova.virt.hardware [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.037442] env[61728]: DEBUG nova.virt.hardware [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.037607] env[61728]: DEBUG nova.virt.hardware [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.037777] env[61728]: DEBUG nova.virt.hardware [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.038719] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8794a84b-990c-494a-9c12-f2931d1602d8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.048754] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5cab4c5-384e-4919-9b59-6c9ca64e251b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.114900] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525995b7-bb93-af31-91fa-ab6adc58cb82, 'name': SearchDatastore_Task, 'duration_secs': 0.011457} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.114900] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.114900] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 788.114900] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.114900] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.115595] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 788.115595] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c2971b4-cf9d-46e1-8d4e-4703e8fae459 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.126155] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 788.126387] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 788.127196] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c8494ce-e3ba-4281-87b3-585f3dac80e4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.133810] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 788.133810] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5233d273-ae2d-da3b-95b0-ce9c926fcb03" [ 788.133810] env[61728]: _type = "Task" [ 788.133810] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.140086] env[61728]: DEBUG nova.compute.manager [req-942ecad1-f38b-419b-9292-20de3fc70f18 req-35ee80c4-db45-405d-9154-e92372b8f8b3 service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Received event network-changed-0e719ade-f327-4ddb-8db5-8beda823ed1b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 788.140086] env[61728]: DEBUG nova.compute.manager [req-942ecad1-f38b-419b-9292-20de3fc70f18 req-35ee80c4-db45-405d-9154-e92372b8f8b3 service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Refreshing instance network info cache due to event network-changed-0e719ade-f327-4ddb-8db5-8beda823ed1b. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 788.140259] env[61728]: DEBUG oslo_concurrency.lockutils [req-942ecad1-f38b-419b-9292-20de3fc70f18 req-35ee80c4-db45-405d-9154-e92372b8f8b3 service nova] Acquiring lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.140395] env[61728]: DEBUG oslo_concurrency.lockutils [req-942ecad1-f38b-419b-9292-20de3fc70f18 req-35ee80c4-db45-405d-9154-e92372b8f8b3 service nova] Acquired lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.140550] env[61728]: DEBUG nova.network.neutron [req-942ecad1-f38b-419b-9292-20de3fc70f18 req-35ee80c4-db45-405d-9154-e92372b8f8b3 service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Refreshing network info cache for port 0e719ade-f327-4ddb-8db5-8beda823ed1b {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 788.148520] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5233d273-ae2d-da3b-95b0-ce9c926fcb03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.206079] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.230s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.208794] env[61728]: DEBUG oslo_concurrency.lockutils [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 43.250s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.209060] env[61728]: DEBUG nova.objects.instance [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lazy-loading 'resources' on Instance uuid 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 788.339392] env[61728]: DEBUG oslo_concurrency.lockutils [req-0ef67276-3a07-445d-ae1f-776b5b95c853 req-7b7bcbd4-8f62-443f-a8f5-484a446037c1 service nova] Releasing lock "refresh_cache-a716cf5a-4482-45cb-96ec-b8c38bc3e742" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.456249] env[61728]: DEBUG nova.network.neutron [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Successfully updated port: 920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 788.647316] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5233d273-ae2d-da3b-95b0-ce9c926fcb03, 'name': SearchDatastore_Task, 'duration_secs': 0.0136} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.648689] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01acfe58-4626-4406-99b0-afffd1e48192 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.657812] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 788.657812] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a97043-9039-db36-fa6f-f21b014a1ec3" [ 788.657812] env[61728]: _type = "Task" [ 788.657812] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.660067] env[61728]: DEBUG nova.compute.manager [req-11123aac-6960-4c57-8382-14c04e9a41d5 req-2d34066f-443a-4fac-a176-24dedebd5d25 service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Received event network-vif-plugged-920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 788.660310] env[61728]: DEBUG oslo_concurrency.lockutils [req-11123aac-6960-4c57-8382-14c04e9a41d5 req-2d34066f-443a-4fac-a176-24dedebd5d25 service nova] Acquiring lock "b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.660531] env[61728]: DEBUG oslo_concurrency.lockutils [req-11123aac-6960-4c57-8382-14c04e9a41d5 req-2d34066f-443a-4fac-a176-24dedebd5d25 service nova] Lock "b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.660705] env[61728]: DEBUG oslo_concurrency.lockutils [req-11123aac-6960-4c57-8382-14c04e9a41d5 req-2d34066f-443a-4fac-a176-24dedebd5d25 service nova] Lock "b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.660880] env[61728]: DEBUG nova.compute.manager [req-11123aac-6960-4c57-8382-14c04e9a41d5 req-2d34066f-443a-4fac-a176-24dedebd5d25 service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] No waiting events found dispatching network-vif-plugged-920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 788.661071] env[61728]: WARNING nova.compute.manager [req-11123aac-6960-4c57-8382-14c04e9a41d5 req-2d34066f-443a-4fac-a176-24dedebd5d25 service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Received unexpected event network-vif-plugged-920c9509-d0d8-464d-a5e2-bef1ba4f5637 for instance with vm_state building and task_state spawning. [ 788.671275] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a97043-9039-db36-fa6f-f21b014a1ec3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.728169] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8f109743-121c-44a9-8e98-6e695f224fc6 tempest-ServerActionsV293TestJSON-156422642 tempest-ServerActionsV293TestJSON-156422642-project-member] Lock "0f75629b-bdec-4593-aceb-cd0478972ccc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 46.847s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.943474] env[61728]: DEBUG nova.network.neutron [req-942ecad1-f38b-419b-9292-20de3fc70f18 req-35ee80c4-db45-405d-9154-e92372b8f8b3 service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Updated VIF entry in instance network info cache for port 0e719ade-f327-4ddb-8db5-8beda823ed1b. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 788.943474] env[61728]: DEBUG nova.network.neutron [req-942ecad1-f38b-419b-9292-20de3fc70f18 req-35ee80c4-db45-405d-9154-e92372b8f8b3 service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Updating instance_info_cache with network_info: [{"id": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "address": "fa:16:3e:44:e0:e6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e719ade-f3", "ovs_interfaceid": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.959045] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquiring lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.959203] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquired lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.959368] env[61728]: DEBUG nova.network.neutron [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 789.171336] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a97043-9039-db36-fa6f-f21b014a1ec3, 'name': SearchDatastore_Task, 'duration_secs': 0.030035} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.174411] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.174745] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a716cf5a-4482-45cb-96ec-b8c38bc3e742/a716cf5a-4482-45cb-96ec-b8c38bc3e742.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 789.175204] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5dab6c5e-d7fd-46a2-971f-2a9e112b5e4d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.184088] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 789.184088] env[61728]: value = "task-464171" [ 789.184088] env[61728]: _type = "Task" [ 789.184088] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.197774] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464171, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.299356] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c728ac5d-eee3-4ea3-bab7-dba4903d2145 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.307410] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6efc2c53-69ce-4bbd-be49-cf62c1bd3f16 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.340098] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edbc61b0-2f75-44e6-bf56-55616c200c77 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.348528] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-271c48d4-4298-47af-b7ac-93085f74e7a4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.362963] env[61728]: DEBUG nova.compute.provider_tree [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.411702] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "6df7b619-8cc4-4dd9-8596-22dc83234a8e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.412013] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "6df7b619-8cc4-4dd9-8596-22dc83234a8e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.412244] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "6df7b619-8cc4-4dd9-8596-22dc83234a8e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.412433] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "6df7b619-8cc4-4dd9-8596-22dc83234a8e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.412614] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "6df7b619-8cc4-4dd9-8596-22dc83234a8e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.414757] env[61728]: INFO nova.compute.manager [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Terminating instance [ 789.416771] env[61728]: DEBUG nova.compute.manager [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 789.416971] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 789.417793] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b549885-b801-4b64-a09a-32a1c94d5dae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.426289] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 789.426485] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93cae2aa-2c45-44d4-bd93-68246f50a0f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.433640] env[61728]: DEBUG oslo_vmware.api [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 789.433640] env[61728]: value = "task-464172" [ 789.433640] env[61728]: _type = "Task" [ 789.433640] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.445090] env[61728]: DEBUG oslo_vmware.api [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464172, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.445578] env[61728]: DEBUG oslo_concurrency.lockutils [req-942ecad1-f38b-419b-9292-20de3fc70f18 req-35ee80c4-db45-405d-9154-e92372b8f8b3 service nova] Releasing lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.510418] env[61728]: DEBUG nova.network.neutron [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 789.699413] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464171, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.707998] env[61728]: DEBUG nova.network.neutron [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Updating instance_info_cache with network_info: [{"id": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "address": "fa:16:3e:c7:4f:2f", "network": {"id": "15423738-1cd1-4170-9505-8cd6b2a672ca", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1314838493-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f742e5f95a3f4ef8b407894687866958", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76e60ff4-204c-4f48-bd0e-2d5fa0a812ef", "external-id": "nsx-vlan-transportzone-854", "segmentation_id": 854, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920c9509-d0", "ovs_interfaceid": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.866024] env[61728]: DEBUG nova.scheduler.client.report [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.950448] env[61728]: DEBUG oslo_vmware.api [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464172, 'name': PowerOffVM_Task, 'duration_secs': 0.198617} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.951346] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 789.952106] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 789.952206] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f218867-676e-4cb6-b41e-8b1d2fa64826 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.039014] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 790.039328] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 790.040387] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Deleting the datastore file [datastore1] 6df7b619-8cc4-4dd9-8596-22dc83234a8e {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 790.040707] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-576d88ae-49c6-464b-94ef-0643157229c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.048889] env[61728]: DEBUG oslo_vmware.api [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 790.048889] env[61728]: value = "task-464174" [ 790.048889] env[61728]: _type = "Task" [ 790.048889] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.061540] env[61728]: DEBUG oslo_vmware.api [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464174, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.200835] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464171, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.998009} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.201104] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a716cf5a-4482-45cb-96ec-b8c38bc3e742/a716cf5a-4482-45cb-96ec-b8c38bc3e742.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 790.201323] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 790.201592] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4a016f7e-0bd7-4a1e-9e64-5b70887c9bba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.210358] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 790.210358] env[61728]: value = "task-464175" [ 790.210358] env[61728]: _type = "Task" [ 790.210358] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.210554] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Releasing lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.210820] env[61728]: DEBUG nova.compute.manager [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Instance network_info: |[{"id": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "address": "fa:16:3e:c7:4f:2f", "network": {"id": "15423738-1cd1-4170-9505-8cd6b2a672ca", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1314838493-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f742e5f95a3f4ef8b407894687866958", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76e60ff4-204c-4f48-bd0e-2d5fa0a812ef", "external-id": "nsx-vlan-transportzone-854", "segmentation_id": 854, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920c9509-d0", "ovs_interfaceid": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 790.211364] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c7:4f:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '76e60ff4-204c-4f48-bd0e-2d5fa0a812ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '920c9509-d0d8-464d-a5e2-bef1ba4f5637', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 790.221835] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Creating folder: Project (f742e5f95a3f4ef8b407894687866958). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 790.225893] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-be39d812-3b14-4e27-9192-6a4f5c45cc28 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.236783] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464175, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.239212] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Created folder: Project (f742e5f95a3f4ef8b407894687866958) in parent group-v121913. [ 790.239471] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Creating folder: Instances. Parent ref: group-v122102. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 790.239664] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-92391211-2eaf-4705-9abd-1b54176e54b7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.251716] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Created folder: Instances in parent group-v122102. [ 790.251999] env[61728]: DEBUG oslo.service.loopingcall [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.252422] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 790.252742] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b059e60b-5aa0-49a6-aa1a-a52888ad8621 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.278062] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 790.278062] env[61728]: value = "task-464178" [ 790.278062] env[61728]: _type = "Task" [ 790.278062] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.287269] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464178, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.372832] env[61728]: DEBUG oslo_concurrency.lockutils [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.164s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.375473] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 44.843s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.401869] env[61728]: INFO nova.scheduler.client.report [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Deleted allocations for instance 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48 [ 790.561167] env[61728]: DEBUG oslo_vmware.api [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464174, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.388722} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.561568] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 790.561682] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 790.561896] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 790.562135] env[61728]: INFO nova.compute.manager [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 790.562392] env[61728]: DEBUG oslo.service.loopingcall [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.562605] env[61728]: DEBUG nova.compute.manager [-] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 790.562703] env[61728]: DEBUG nova.network.neutron [-] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 790.725804] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464175, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.124672} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.726194] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 790.726996] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e752b0-1849-4090-abb4-c462569d1354 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.759259] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] a716cf5a-4482-45cb-96ec-b8c38bc3e742/a716cf5a-4482-45cb-96ec-b8c38bc3e742.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 790.759601] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40876500-a0e0-4e19-8f20-7259a91c77d2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.785067] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 790.785067] env[61728]: value = "task-464179" [ 790.785067] env[61728]: _type = "Task" [ 790.785067] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.792314] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464178, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.798599] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464179, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.881728] env[61728]: INFO nova.compute.claims [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.894140] env[61728]: DEBUG nova.compute.manager [req-afbc103c-94b1-4ec0-9bc4-ca337e37a8ff req-40cce42a-b52d-4a48-9ec6-5de2308bdd7d service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Received event network-changed-920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 790.894140] env[61728]: DEBUG nova.compute.manager [req-afbc103c-94b1-4ec0-9bc4-ca337e37a8ff req-40cce42a-b52d-4a48-9ec6-5de2308bdd7d service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Refreshing instance network info cache due to event network-changed-920c9509-d0d8-464d-a5e2-bef1ba4f5637. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 790.894140] env[61728]: DEBUG oslo_concurrency.lockutils [req-afbc103c-94b1-4ec0-9bc4-ca337e37a8ff req-40cce42a-b52d-4a48-9ec6-5de2308bdd7d service nova] Acquiring lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.897499] env[61728]: DEBUG oslo_concurrency.lockutils [req-afbc103c-94b1-4ec0-9bc4-ca337e37a8ff req-40cce42a-b52d-4a48-9ec6-5de2308bdd7d service nova] Acquired lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.898056] env[61728]: DEBUG nova.network.neutron [req-afbc103c-94b1-4ec0-9bc4-ca337e37a8ff req-40cce42a-b52d-4a48-9ec6-5de2308bdd7d service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Refreshing network info cache for port 920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 790.913825] env[61728]: DEBUG oslo_concurrency.lockutils [None req-98dd08dc-cf57-49f5-9471-7102fd3be2fe tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 48.946s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.914333] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 45.697s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.914616] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.914815] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.915019] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.922246] env[61728]: INFO nova.compute.manager [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Terminating instance [ 790.925168] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.925356] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquired lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.925534] env[61728]: DEBUG nova.network.neutron [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.293719] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464178, 'name': CreateVM_Task, 'duration_secs': 0.718912} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.294387] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 791.295329] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.295329] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.295532] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 791.295793] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7ed7b15-e16e-4869-8341-3ad2394d818b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.302113] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464179, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.303905] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 791.303905] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b631fa-3d7b-3bcb-45a8-3d2c629e7799" [ 791.303905] env[61728]: _type = "Task" [ 791.303905] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.312704] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b631fa-3d7b-3bcb-45a8-3d2c629e7799, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.378457] env[61728]: DEBUG nova.network.neutron [-] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.404346] env[61728]: INFO nova.compute.resource_tracker [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating resource usage from migration adc30b71-8803-41e0-a72a-09cb8bee068c [ 791.428554] env[61728]: DEBUG nova.compute.utils [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Can not refresh info_cache because instance was not found {{(pid=61728) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 791.450898] env[61728]: DEBUG nova.network.neutron [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 791.586885] env[61728]: DEBUG nova.network.neutron [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.678156] env[61728]: DEBUG nova.network.neutron [req-afbc103c-94b1-4ec0-9bc4-ca337e37a8ff req-40cce42a-b52d-4a48-9ec6-5de2308bdd7d service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Updated VIF entry in instance network info cache for port 920c9509-d0d8-464d-a5e2-bef1ba4f5637. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 791.678485] env[61728]: DEBUG nova.network.neutron [req-afbc103c-94b1-4ec0-9bc4-ca337e37a8ff req-40cce42a-b52d-4a48-9ec6-5de2308bdd7d service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Updating instance_info_cache with network_info: [{"id": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "address": "fa:16:3e:c7:4f:2f", "network": {"id": "15423738-1cd1-4170-9505-8cd6b2a672ca", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1314838493-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f742e5f95a3f4ef8b407894687866958", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76e60ff4-204c-4f48-bd0e-2d5fa0a812ef", "external-id": "nsx-vlan-transportzone-854", "segmentation_id": 854, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920c9509-d0", "ovs_interfaceid": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.801466] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464179, 'name': ReconfigVM_Task, 'duration_secs': 0.566086} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.801785] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Reconfigured VM instance instance-0000003e to attach disk [datastore1] a716cf5a-4482-45cb-96ec-b8c38bc3e742/a716cf5a-4482-45cb-96ec-b8c38bc3e742.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.804199] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ba9daae1-6e84-422c-998d-a3de92156846 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.819885] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b631fa-3d7b-3bcb-45a8-3d2c629e7799, 'name': SearchDatastore_Task, 'duration_secs': 0.031868} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.821391] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.821671] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 791.821927] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.822094] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.822285] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 791.822631] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 791.822631] env[61728]: value = "task-464180" [ 791.822631] env[61728]: _type = "Task" [ 791.822631] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.822844] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-130549c5-8319-48a7-b46a-48085c68d9a9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.839092] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464180, 'name': Rename_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.842635] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 791.842946] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 791.843783] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d6a697e-a784-41c3-86bf-6b807c9f0bac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.853625] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 791.853625] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5283d016-b928-f1f6-1e98-06defbf000ba" [ 791.853625] env[61728]: _type = "Task" [ 791.853625] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.867653] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5283d016-b928-f1f6-1e98-06defbf000ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.882771] env[61728]: INFO nova.compute.manager [-] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Took 1.32 seconds to deallocate network for instance. [ 791.999273] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499c7545-8961-41d8-aef9-055d16396363 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.007544] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e031114e-c450-40ed-bb82-33416ccef890 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.040728] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fca36c8-4e4c-447d-bcb5-83eaf5b656df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.052133] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7fdf47f-58c4-4dff-a315-907ad694f43a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.069098] env[61728]: DEBUG nova.compute.provider_tree [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.089265] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Releasing lock "refresh_cache-3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.089741] env[61728]: DEBUG nova.compute.manager [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 792.089942] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 792.090428] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8147801-6f74-48af-8b70-6443adb1d7f3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.104480] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-063f1679-9f25-45fb-9383-568f61f9bc23 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.147369] env[61728]: WARNING nova.virt.vmwareapi.vmops [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48 could not be found. [ 792.147574] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 792.147808] env[61728]: INFO nova.compute.manager [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Took 0.06 seconds to destroy the instance on the hypervisor. [ 792.148091] env[61728]: DEBUG oslo.service.loopingcall [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.148742] env[61728]: DEBUG nova.compute.manager [-] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 792.148742] env[61728]: DEBUG nova.network.neutron [-] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 792.171646] env[61728]: DEBUG nova.network.neutron [-] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 792.181414] env[61728]: DEBUG oslo_concurrency.lockutils [req-afbc103c-94b1-4ec0-9bc4-ca337e37a8ff req-40cce42a-b52d-4a48-9ec6-5de2308bdd7d service nova] Releasing lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.339178] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464180, 'name': Rename_Task, 'duration_secs': 0.148227} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.339942] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 792.342324] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-465f5d0f-4e96-472c-86f1-c7b287001b20 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.351193] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 792.351193] env[61728]: value = "task-464181" [ 792.351193] env[61728]: _type = "Task" [ 792.351193] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.361903] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464181, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.370103] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5283d016-b928-f1f6-1e98-06defbf000ba, 'name': SearchDatastore_Task, 'duration_secs': 0.012916} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.370103] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81044497-fde0-4a16-9314-4f85fe740f15 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.375496] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 792.375496] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528f2ae2-2134-1b6a-f913-04e0c1e6e141" [ 792.375496] env[61728]: _type = "Task" [ 792.375496] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.380072] env[61728]: DEBUG oslo_concurrency.lockutils [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "0fb1192e-99f1-4469-b196-60df7eab8185" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.380072] env[61728]: DEBUG oslo_concurrency.lockutils [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "0fb1192e-99f1-4469-b196-60df7eab8185" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.380072] env[61728]: DEBUG oslo_concurrency.lockutils [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "0fb1192e-99f1-4469-b196-60df7eab8185-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.380072] env[61728]: DEBUG oslo_concurrency.lockutils [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "0fb1192e-99f1-4469-b196-60df7eab8185-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.380072] env[61728]: DEBUG oslo_concurrency.lockutils [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "0fb1192e-99f1-4469-b196-60df7eab8185-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.383112] env[61728]: INFO nova.compute.manager [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Terminating instance [ 792.389181] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.389537] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528f2ae2-2134-1b6a-f913-04e0c1e6e141, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.390371] env[61728]: DEBUG nova.compute.manager [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 792.390803] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 792.391746] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d552da-b551-40fb-a188-9a5abd020e76 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.400475] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 792.400782] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64619354-47d2-4a4f-883b-737839338afd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.408453] env[61728]: DEBUG oslo_vmware.api [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 792.408453] env[61728]: value = "task-464182" [ 792.408453] env[61728]: _type = "Task" [ 792.408453] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.422707] env[61728]: DEBUG oslo_vmware.api [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464182, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.572046] env[61728]: DEBUG nova.scheduler.client.report [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.676161] env[61728]: DEBUG nova.network.neutron [-] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.869199] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464181, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.897589] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528f2ae2-2134-1b6a-f913-04e0c1e6e141, 'name': SearchDatastore_Task, 'duration_secs': 0.016972} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.897869] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.898280] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4/b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 792.898568] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cded8990-1e3a-497f-bda4-06841583dd46 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.907634] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 792.907634] env[61728]: value = "task-464183" [ 792.907634] env[61728]: _type = "Task" [ 792.907634] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.921046] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464183, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.924212] env[61728]: DEBUG oslo_vmware.api [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464182, 'name': PowerOffVM_Task, 'duration_secs': 0.315919} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.927371] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 792.927371] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 792.927371] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7d3d74bc-f9fa-4b25-b8b8-0ec6452b071b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.997816] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 792.998120] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 792.998348] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Deleting the datastore file [datastore1] 0fb1192e-99f1-4469-b196-60df7eab8185 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 792.998662] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11a5de7f-3c43-48bc-b908-b135a739a009 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.008294] env[61728]: DEBUG oslo_vmware.api [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for the task: (returnval){ [ 793.008294] env[61728]: value = "task-464185" [ 793.008294] env[61728]: _type = "Task" [ 793.008294] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.019711] env[61728]: DEBUG oslo_vmware.api [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464185, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.023982] env[61728]: DEBUG nova.compute.manager [req-e28a4bc8-281e-43da-8367-387e42375dea req-d0e271f6-f88d-4fa6-9a96-5c89b84062fe service nova] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Received event network-vif-deleted-551e9af7-6eba-46a7-89a6-27f48212864d {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 793.078312] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.702s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.078312] env[61728]: INFO nova.compute.manager [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Migrating [ 793.085014] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 45.484s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.086675] env[61728]: INFO nova.compute.claims [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.183118] env[61728]: INFO nova.compute.manager [-] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Took 1.03 seconds to deallocate network for instance. [ 793.364649] env[61728]: DEBUG oslo_vmware.api [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464181, 'name': PowerOnVM_Task, 'duration_secs': 0.531266} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.365109] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 793.365349] env[61728]: INFO nova.compute.manager [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Took 12.07 seconds to spawn the instance on the hypervisor. [ 793.365571] env[61728]: DEBUG nova.compute.manager [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 793.366438] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00efbb61-01ed-40a4-ab98-46245c8283c8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.425715] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464183, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.526246] env[61728]: DEBUG oslo_vmware.api [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Task: {'id': task-464185, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.420498} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.526500] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 793.526984] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 793.526984] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 793.527092] env[61728]: INFO nova.compute.manager [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Took 1.14 seconds to destroy the instance on the hypervisor. [ 793.527318] env[61728]: DEBUG oslo.service.loopingcall [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 793.527525] env[61728]: DEBUG nova.compute.manager [-] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 793.527624] env[61728]: DEBUG nova.network.neutron [-] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 793.599825] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.599973] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.600323] env[61728]: DEBUG nova.network.neutron [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 793.692893] env[61728]: INFO nova.compute.manager [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Instance disappeared during terminate [ 793.692893] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dffedbd6-a5b0-4760-95fe-c814fa6b2274 tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.777s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.723339] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Acquiring lock "174607a1-9bc8-4e07-8993-7f0bb0f308e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.724726] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Lock "174607a1-9bc8-4e07-8993-7f0bb0f308e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.888879] env[61728]: INFO nova.compute.manager [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Took 60.06 seconds to build instance. [ 793.921072] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464183, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.601137} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.921647] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4/b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 793.921647] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 793.921858] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f6728cf9-3000-4e1d-ac09-65d5d7140073 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.929351] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 793.929351] env[61728]: value = "task-464186" [ 793.929351] env[61728]: _type = "Task" [ 793.929351] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.938076] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464186, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.226515] env[61728]: DEBUG nova.compute.manager [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.296578] env[61728]: DEBUG nova.network.neutron [-] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.358069] env[61728]: DEBUG nova.network.neutron [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating instance_info_cache with network_info: [{"id": "cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd", "address": "fa:16:3e:fc:46:55", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc2e7ae1-f4", "ovs_interfaceid": "cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.390123] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3447dbc3-a6db-4fca-90bc-763aa11af86d tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.784s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.442951] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464186, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.293794} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.445802] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 794.447286] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b6dde27-4c06-477c-8ccf-48d73451d25d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.471925] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4/b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 794.475270] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f205300a-d79c-4b96-965e-e078c4744c77 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.500568] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 794.500568] env[61728]: value = "task-464187" [ 794.500568] env[61728]: _type = "Task" [ 794.500568] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.523391] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464187, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.719425] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f0d2233-29a0-4a2f-b277-72fdf1f20a75 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.727269] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-143fc5cb-0cb8-4ed5-a6cd-ea08df860249 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.765259] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de6f1c2-1a11-45e5-8b1e-4853ad098140 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.773516] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1058c995-89ce-47a0-b3b1-06a97ee20f12 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.790537] env[61728]: DEBUG nova.compute.provider_tree [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.794208] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.799745] env[61728]: INFO nova.compute.manager [-] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Took 1.27 seconds to deallocate network for instance. [ 794.861527] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.007056] env[61728]: DEBUG oslo_concurrency.lockutils [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.007337] env[61728]: DEBUG oslo_concurrency.lockutils [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.008288] env[61728]: DEBUG oslo_concurrency.lockutils [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.008288] env[61728]: DEBUG oslo_concurrency.lockutils [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.008288] env[61728]: DEBUG oslo_concurrency.lockutils [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.010485] env[61728]: INFO nova.compute.manager [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Terminating instance [ 795.015978] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464187, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.016577] env[61728]: DEBUG nova.compute.manager [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 795.016777] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 795.017637] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d177dc-bb2b-4d1c-9a78-131116264a78 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.024915] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 795.024915] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-332f21ae-4f98-4b36-b521-ced2613edfe2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.031426] env[61728]: DEBUG oslo_vmware.api [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 795.031426] env[61728]: value = "task-464188" [ 795.031426] env[61728]: _type = "Task" [ 795.031426] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.039425] env[61728]: DEBUG oslo_vmware.api [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464188, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.064218] env[61728]: DEBUG nova.compute.manager [req-cd663505-1737-4192-870c-4337519fb795 req-eb75e3f7-5f24-49eb-ad04-08d26a256d93 service nova] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Received event network-vif-deleted-0b046d2a-a7b2-47d4-b5e9-da3b3e2807d3 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 795.295326] env[61728]: DEBUG nova.scheduler.client.report [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.309021] env[61728]: DEBUG oslo_concurrency.lockutils [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.512603] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464187, 'name': ReconfigVM_Task, 'duration_secs': 0.614883} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.512906] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Reconfigured VM instance instance-0000003f to attach disk [datastore1] b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4/b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.513575] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2790c1e9-897e-4df1-b999-24f265b1053c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.521254] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 795.521254] env[61728]: value = "task-464189" [ 795.521254] env[61728]: _type = "Task" [ 795.521254] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.530593] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464189, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.539372] env[61728]: DEBUG oslo_vmware.api [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464188, 'name': PowerOffVM_Task, 'duration_secs': 0.252544} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.539667] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 795.539848] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 795.540104] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dbdfaa43-341d-49f9-a256-28b23f7d4809 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.647286] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 795.647527] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 795.649047] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Deleting the datastore file [datastore1] a716cf5a-4482-45cb-96ec-b8c38bc3e742 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 795.649047] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28776a36-8dfc-4e3d-812e-c2b3ebe8d604 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.655076] env[61728]: DEBUG oslo_vmware.api [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for the task: (returnval){ [ 795.655076] env[61728]: value = "task-464191" [ 795.655076] env[61728]: _type = "Task" [ 795.655076] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.663406] env[61728]: DEBUG oslo_vmware.api [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464191, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.800294] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.715s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.800965] env[61728]: DEBUG nova.compute.manager [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 795.804154] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 45.707s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.804154] env[61728]: DEBUG nova.objects.instance [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Lazy-loading 'resources' on Instance uuid 8c4b7227-1a01-4400-beb8-2df27e17e329 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.032891] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464189, 'name': Rename_Task, 'duration_secs': 0.1413} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.033221] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 796.033480] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c3c57f3a-3976-471a-949b-5072bdff4b14 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.040713] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 796.040713] env[61728]: value = "task-464192" [ 796.040713] env[61728]: _type = "Task" [ 796.040713] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.049557] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464192, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.165495] env[61728]: DEBUG oslo_vmware.api [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Task: {'id': task-464191, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138027} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.165774] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 796.165970] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 796.166170] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 796.166354] env[61728]: INFO nova.compute.manager [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Took 1.15 seconds to destroy the instance on the hypervisor. [ 796.166601] env[61728]: DEBUG oslo.service.loopingcall [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.166859] env[61728]: DEBUG nova.compute.manager [-] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 796.166959] env[61728]: DEBUG nova.network.neutron [-] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 796.307090] env[61728]: DEBUG nova.compute.utils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 796.311253] env[61728]: DEBUG nova.compute.manager [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 796.311429] env[61728]: DEBUG nova.network.neutron [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 796.375090] env[61728]: DEBUG nova.policy [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '091abd20ebfe450b9283f5c7df1c7993', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c78292171414530a1caba3dfd0f620f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 796.379924] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d076029d-f507-4dd7-b484-721d53f120b1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.406363] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating instance 'f2692e22-6c14-4bb3-a080-607f0731105d' progress to 0 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 796.550690] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464192, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.798851] env[61728]: DEBUG nova.network.neutron [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Successfully created port: 4758105c-971a-4c28-a179-7d9e20f05173 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 796.812481] env[61728]: DEBUG nova.compute.manager [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 796.918695] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 796.919331] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2f36363-1a5f-4e27-af58-615d238e4c38 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.926738] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 796.926738] env[61728]: value = "task-464193" [ 796.926738] env[61728]: _type = "Task" [ 796.926738] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.935988] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464193, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.967142] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fede9ade-f975-4450-a6bb-928160a0b187 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.976143] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf7b497-b767-412f-b99c-a0039f882ea3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.014244] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29641e8c-c892-4bc5-bd55-54200f089359 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.021653] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37209f41-95ef-4934-873a-57bc45f3cc2c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.036809] env[61728]: DEBUG nova.compute.provider_tree [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.053392] env[61728]: DEBUG oslo_vmware.api [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464192, 'name': PowerOnVM_Task, 'duration_secs': 1.011039} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.053726] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 797.053954] env[61728]: INFO nova.compute.manager [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Took 9.06 seconds to spawn the instance on the hypervisor. [ 797.054197] env[61728]: DEBUG nova.compute.manager [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 797.055060] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e669f9-7028-40f5-a0ac-321a04d1ab39 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.180450] env[61728]: DEBUG nova.compute.manager [req-6d444ddb-65c6-458e-adff-6c024f16fc43 req-16fe803d-2a86-4598-8dfe-acc3e3d4ef1f service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Received event network-vif-deleted-c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 797.180450] env[61728]: INFO nova.compute.manager [req-6d444ddb-65c6-458e-adff-6c024f16fc43 req-16fe803d-2a86-4598-8dfe-acc3e3d4ef1f service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Neutron deleted interface c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49; detaching it from the instance and deleting it from the info cache [ 797.180450] env[61728]: DEBUG nova.network.neutron [req-6d444ddb-65c6-458e-adff-6c024f16fc43 req-16fe803d-2a86-4598-8dfe-acc3e3d4ef1f service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Updating instance_info_cache with network_info: [{"id": "60b8853e-0428-415e-99cb-0a8541ea7a64", "address": "fa:16:3e:06:1b:f3", "network": {"id": "4a8d636f-6cc7-45af-acde-fa1c06e488b3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1818596159", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.164", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "037e1941b7964a2b94d5b0f5c280956e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60b8853e-04", "ovs_interfaceid": "60b8853e-0428-415e-99cb-0a8541ea7a64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.268364] env[61728]: DEBUG nova.network.neutron [-] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.436665] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464193, 'name': PowerOffVM_Task, 'duration_secs': 0.179736} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.436959] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 797.437169] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating instance 'f2692e22-6c14-4bb3-a080-607f0731105d' progress to 17 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 797.540242] env[61728]: DEBUG nova.scheduler.client.report [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.571813] env[61728]: INFO nova.compute.manager [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Took 55.46 seconds to build instance. [ 797.683634] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-82db7d3b-1eb1-4137-bc23-00e9fd33da1c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.693478] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1512e6e4-572a-45f1-825e-b50f581d8b1c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.729054] env[61728]: DEBUG nova.compute.manager [req-6d444ddb-65c6-458e-adff-6c024f16fc43 req-16fe803d-2a86-4598-8dfe-acc3e3d4ef1f service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Detach interface failed, port_id=c8a2ce4f-4dd6-40e4-b4df-42c16dfbdb49, reason: Instance a716cf5a-4482-45cb-96ec-b8c38bc3e742 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 797.729323] env[61728]: DEBUG nova.compute.manager [req-6d444ddb-65c6-458e-adff-6c024f16fc43 req-16fe803d-2a86-4598-8dfe-acc3e3d4ef1f service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Received event network-vif-deleted-60b8853e-0428-415e-99cb-0a8541ea7a64 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 797.729505] env[61728]: INFO nova.compute.manager [req-6d444ddb-65c6-458e-adff-6c024f16fc43 req-16fe803d-2a86-4598-8dfe-acc3e3d4ef1f service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Neutron deleted interface 60b8853e-0428-415e-99cb-0a8541ea7a64; detaching it from the instance and deleting it from the info cache [ 797.730491] env[61728]: DEBUG nova.network.neutron [req-6d444ddb-65c6-458e-adff-6c024f16fc43 req-16fe803d-2a86-4598-8dfe-acc3e3d4ef1f service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.771872] env[61728]: INFO nova.compute.manager [-] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Took 1.60 seconds to deallocate network for instance. [ 797.822541] env[61728]: DEBUG nova.compute.manager [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 797.903147] env[61728]: DEBUG nova.virt.hardware [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 797.903417] env[61728]: DEBUG nova.virt.hardware [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 797.903583] env[61728]: DEBUG nova.virt.hardware [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 797.903772] env[61728]: DEBUG nova.virt.hardware [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 797.903927] env[61728]: DEBUG nova.virt.hardware [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 797.904270] env[61728]: DEBUG nova.virt.hardware [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 797.904436] env[61728]: DEBUG nova.virt.hardware [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 797.904606] env[61728]: DEBUG nova.virt.hardware [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 797.904774] env[61728]: DEBUG nova.virt.hardware [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 797.904942] env[61728]: DEBUG nova.virt.hardware [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 797.905166] env[61728]: DEBUG nova.virt.hardware [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 797.906006] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c615b5a1-f42b-4f99-a7d0-4198387b633c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.916607] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a0eae3-889f-4bf1-91aa-b44d5b756155 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.942987] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 797.943262] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 797.943423] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 797.943613] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 797.943774] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 797.943927] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 797.944146] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 797.944307] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 797.944485] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 797.944650] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 797.944829] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 797.949904] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21697b16-d59b-40ab-b200-f4bf7e945c09 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.965111] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 797.965111] env[61728]: value = "task-464194" [ 797.965111] env[61728]: _type = "Task" [ 797.965111] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.973421] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464194, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.047883] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.244s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.050523] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 44.218s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.052309] env[61728]: INFO nova.compute.claims [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.074138] env[61728]: DEBUG oslo_concurrency.lockutils [None req-81463790-1eae-4366-9d10-5b8238a2b987 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Lock "b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.247s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.076036] env[61728]: INFO nova.scheduler.client.report [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Deleted allocations for instance 8c4b7227-1a01-4400-beb8-2df27e17e329 [ 798.233139] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-46bf6384-dbaa-4759-a8f0-6d3a738c0bee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.242864] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de5e0d9-1463-483d-8a4b-0e11b906faa1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.275843] env[61728]: DEBUG nova.compute.manager [req-6d444ddb-65c6-458e-adff-6c024f16fc43 req-16fe803d-2a86-4598-8dfe-acc3e3d4ef1f service nova] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Detach interface failed, port_id=60b8853e-0428-415e-99cb-0a8541ea7a64, reason: Instance a716cf5a-4482-45cb-96ec-b8c38bc3e742 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 798.279096] env[61728]: DEBUG oslo_concurrency.lockutils [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.295718] env[61728]: INFO nova.compute.manager [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Rescuing [ 798.295975] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquiring lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.296149] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquired lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.296347] env[61728]: DEBUG nova.network.neutron [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 798.340288] env[61728]: DEBUG nova.network.neutron [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Successfully updated port: 4758105c-971a-4c28-a179-7d9e20f05173 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 798.475060] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464194, 'name': ReconfigVM_Task, 'duration_secs': 0.204707} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.475449] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating instance 'f2692e22-6c14-4bb3-a080-607f0731105d' progress to 33 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 798.584534] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8de208ce-57bf-4a06-96b3-bdb051533755 tempest-VolumesAssistedSnapshotsTest-952521230 tempest-VolumesAssistedSnapshotsTest-952521230-project-member] Lock "8c4b7227-1a01-4400-beb8-2df27e17e329" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 51.891s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.844957] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "refresh_cache-06109957-2c3b-404e-a43e-dd34ece39096" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.845030] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "refresh_cache-06109957-2c3b-404e-a43e-dd34ece39096" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.845183] env[61728]: DEBUG nova.network.neutron [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 798.982250] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.982778] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.982778] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.982966] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.983538] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.984346] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.984346] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.984346] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.984541] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.984750] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.984971] env[61728]: DEBUG nova.virt.hardware [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.990372] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Reconfiguring VM instance instance-00000035 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 798.991042] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b93c4160-6e85-4c9a-b140-f23595adf157 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.014708] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 799.014708] env[61728]: value = "task-464195" [ 799.014708] env[61728]: _type = "Task" [ 799.014708] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.021392] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464195, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.041516] env[61728]: DEBUG nova.network.neutron [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Updating instance_info_cache with network_info: [{"id": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "address": "fa:16:3e:c7:4f:2f", "network": {"id": "15423738-1cd1-4170-9505-8cd6b2a672ca", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1314838493-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f742e5f95a3f4ef8b407894687866958", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76e60ff4-204c-4f48-bd0e-2d5fa0a812ef", "external-id": "nsx-vlan-transportzone-854", "segmentation_id": 854, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920c9509-d0", "ovs_interfaceid": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.269268] env[61728]: DEBUG nova.compute.manager [req-cecd82fb-2e5c-498c-9311-67af05d93aa3 req-8456b73e-5eff-4eb7-9243-be6e7ac17d6c service nova] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Received event network-vif-plugged-4758105c-971a-4c28-a179-7d9e20f05173 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 799.269940] env[61728]: DEBUG oslo_concurrency.lockutils [req-cecd82fb-2e5c-498c-9311-67af05d93aa3 req-8456b73e-5eff-4eb7-9243-be6e7ac17d6c service nova] Acquiring lock "06109957-2c3b-404e-a43e-dd34ece39096-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.270427] env[61728]: DEBUG oslo_concurrency.lockutils [req-cecd82fb-2e5c-498c-9311-67af05d93aa3 req-8456b73e-5eff-4eb7-9243-be6e7ac17d6c service nova] Lock "06109957-2c3b-404e-a43e-dd34ece39096-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.270833] env[61728]: DEBUG oslo_concurrency.lockutils [req-cecd82fb-2e5c-498c-9311-67af05d93aa3 req-8456b73e-5eff-4eb7-9243-be6e7ac17d6c service nova] Lock "06109957-2c3b-404e-a43e-dd34ece39096-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.272233] env[61728]: DEBUG nova.compute.manager [req-cecd82fb-2e5c-498c-9311-67af05d93aa3 req-8456b73e-5eff-4eb7-9243-be6e7ac17d6c service nova] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] No waiting events found dispatching network-vif-plugged-4758105c-971a-4c28-a179-7d9e20f05173 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 799.272233] env[61728]: WARNING nova.compute.manager [req-cecd82fb-2e5c-498c-9311-67af05d93aa3 req-8456b73e-5eff-4eb7-9243-be6e7ac17d6c service nova] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Received unexpected event network-vif-plugged-4758105c-971a-4c28-a179-7d9e20f05173 for instance with vm_state building and task_state spawning. [ 799.272233] env[61728]: DEBUG nova.compute.manager [req-cecd82fb-2e5c-498c-9311-67af05d93aa3 req-8456b73e-5eff-4eb7-9243-be6e7ac17d6c service nova] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Received event network-changed-4758105c-971a-4c28-a179-7d9e20f05173 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 799.272233] env[61728]: DEBUG nova.compute.manager [req-cecd82fb-2e5c-498c-9311-67af05d93aa3 req-8456b73e-5eff-4eb7-9243-be6e7ac17d6c service nova] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Refreshing instance network info cache due to event network-changed-4758105c-971a-4c28-a179-7d9e20f05173. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 799.272233] env[61728]: DEBUG oslo_concurrency.lockutils [req-cecd82fb-2e5c-498c-9311-67af05d93aa3 req-8456b73e-5eff-4eb7-9243-be6e7ac17d6c service nova] Acquiring lock "refresh_cache-06109957-2c3b-404e-a43e-dd34ece39096" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.419915] env[61728]: DEBUG nova.network.neutron [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 799.531343] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464195, 'name': ReconfigVM_Task, 'duration_secs': 0.170131} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.534174] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Reconfigured VM instance instance-00000035 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 799.535023] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90bc4ca3-8d6a-4746-8703-25e9fd325fe2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.554746] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Releasing lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.561772] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] f2692e22-6c14-4bb3-a080-607f0731105d/f2692e22-6c14-4bb3-a080-607f0731105d.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 799.564540] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c65e2ce-9ec5-44a5-9cac-7fde9cac7dcd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.583707] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 799.583707] env[61728]: value = "task-464196" [ 799.583707] env[61728]: _type = "Task" [ 799.583707] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.601118] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464196, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.634265] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8be2d07-682c-41f2-8b00-db683ce6d628 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.643062] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b345b60b-8d60-4b5e-ba42-22a961a0a19c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.675608] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e60ae6-14a9-49fd-b8e5-7f7518134861 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.679066] env[61728]: DEBUG nova.network.neutron [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Updating instance_info_cache with network_info: [{"id": "4758105c-971a-4c28-a179-7d9e20f05173", "address": "fa:16:3e:ca:0d:84", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4758105c-97", "ovs_interfaceid": "4758105c-971a-4c28-a179-7d9e20f05173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.685902] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-481729cd-e2ca-45e3-af04-7db23250cf0b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.700299] env[61728]: DEBUG nova.compute.provider_tree [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.097579] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464196, 'name': ReconfigVM_Task, 'duration_secs': 0.254429} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.097875] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Reconfigured VM instance instance-00000035 to attach disk [datastore1] f2692e22-6c14-4bb3-a080-607f0731105d/f2692e22-6c14-4bb3-a080-607f0731105d.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.098172] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating instance 'f2692e22-6c14-4bb3-a080-607f0731105d' progress to 50 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 800.110875] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 800.111137] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e3281a95-b8a6-418a-b440-5327ef9fc66c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.117100] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 800.117100] env[61728]: value = "task-464197" [ 800.117100] env[61728]: _type = "Task" [ 800.117100] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.125530] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464197, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.181578] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "refresh_cache-06109957-2c3b-404e-a43e-dd34ece39096" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.181990] env[61728]: DEBUG nova.compute.manager [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Instance network_info: |[{"id": "4758105c-971a-4c28-a179-7d9e20f05173", "address": "fa:16:3e:ca:0d:84", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4758105c-97", "ovs_interfaceid": "4758105c-971a-4c28-a179-7d9e20f05173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 800.183029] env[61728]: DEBUG oslo_concurrency.lockutils [req-cecd82fb-2e5c-498c-9311-67af05d93aa3 req-8456b73e-5eff-4eb7-9243-be6e7ac17d6c service nova] Acquired lock "refresh_cache-06109957-2c3b-404e-a43e-dd34ece39096" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.183029] env[61728]: DEBUG nova.network.neutron [req-cecd82fb-2e5c-498c-9311-67af05d93aa3 req-8456b73e-5eff-4eb7-9243-be6e7ac17d6c service nova] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Refreshing network info cache for port 4758105c-971a-4c28-a179-7d9e20f05173 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 800.183739] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:0d:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4758105c-971a-4c28-a179-7d9e20f05173', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.191136] env[61728]: DEBUG oslo.service.loopingcall [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.192575] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 800.192575] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-697ba3c0-f326-4b5e-a0e2-2de589c40564 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.211169] env[61728]: DEBUG nova.scheduler.client.report [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.221738] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.221738] env[61728]: value = "task-464198" [ 800.221738] env[61728]: _type = "Task" [ 800.221738] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.229565] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464198, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.609908] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078a30d2-47a6-409d-8f7f-2529817b48e6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.638430] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ee8617-bf5b-438b-983c-515ccaa6305d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.651424] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464197, 'name': PowerOffVM_Task, 'duration_secs': 0.205275} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.666961] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 800.667113] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating instance 'f2692e22-6c14-4bb3-a080-607f0731105d' progress to 67 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 800.671857] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0664ef1e-a13c-4fa0-bcd9-5e4ac74e9881 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.697743] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0113497-f8b8-4da2-8ed4-41e08fb2c7b9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.717943] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.667s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.718554] env[61728]: DEBUG nova.compute.manager [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 800.725288] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 45.929s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.725288] env[61728]: DEBUG nova.objects.instance [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lazy-loading 'resources' on Instance uuid b563e521-0e0e-45ea-bd5a-056285ea35b9 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 800.739015] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464198, 'name': CreateVM_Task, 'duration_secs': 0.380175} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.739704] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 800.740775] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.740775] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.740908] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 800.741682] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84b51d37-5334-47ce-ada9-5d6f7f6f465a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.748183] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 800.748183] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c194bda7-7fd2-4c56-96c8-a7ac15ace383 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.751534] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 800.751534] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52585edd-b167-29ed-ffdf-ac11c95afba1" [ 800.751534] env[61728]: _type = "Task" [ 800.751534] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.756826] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 800.756826] env[61728]: value = "task-464199" [ 800.756826] env[61728]: _type = "Task" [ 800.756826] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.764624] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52585edd-b167-29ed-ffdf-ac11c95afba1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.769687] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 800.769989] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 800.770306] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.770584] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.770817] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 800.771358] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a53800dd-3e87-4d22-8b87-63a7be5c8ab2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.782482] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 800.782676] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 800.783432] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d5be052-6476-4c11-bbae-c40b715e302c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.788516] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 800.788516] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f158bb-deea-e3fd-9fef-c27039385c0c" [ 800.788516] env[61728]: _type = "Task" [ 800.788516] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.795863] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f158bb-deea-e3fd-9fef-c27039385c0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.064035] env[61728]: DEBUG nova.network.neutron [req-cecd82fb-2e5c-498c-9311-67af05d93aa3 req-8456b73e-5eff-4eb7-9243-be6e7ac17d6c service nova] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Updated VIF entry in instance network info cache for port 4758105c-971a-4c28-a179-7d9e20f05173. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 801.064524] env[61728]: DEBUG nova.network.neutron [req-cecd82fb-2e5c-498c-9311-67af05d93aa3 req-8456b73e-5eff-4eb7-9243-be6e7ac17d6c service nova] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Updating instance_info_cache with network_info: [{"id": "4758105c-971a-4c28-a179-7d9e20f05173", "address": "fa:16:3e:ca:0d:84", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4758105c-97", "ovs_interfaceid": "4758105c-971a-4c28-a179-7d9e20f05173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.228901] env[61728]: DEBUG nova.compute.utils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 801.230471] env[61728]: DEBUG nova.compute.manager [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 801.230645] env[61728]: DEBUG nova.network.neutron [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 801.237775] env[61728]: DEBUG nova.network.neutron [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Port cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd binding to destination host cpu-1 is already ACTIVE {{(pid=61728) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 801.264287] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52585edd-b167-29ed-ffdf-ac11c95afba1, 'name': SearchDatastore_Task, 'duration_secs': 0.02151} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.266944] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.267303] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 801.267525] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.300078] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f158bb-deea-e3fd-9fef-c27039385c0c, 'name': SearchDatastore_Task, 'duration_secs': 0.035964} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.301750] env[61728]: DEBUG nova.policy [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd32b5118a94042f797dea1fdf6bd05b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08367aaceba548fe93faaedf6371817d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 801.305765] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-defa5265-e1e4-4a95-a64d-713c6e9d0b65 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.311048] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 801.311048] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b572e4-36d4-6335-ca88-d9d797722619" [ 801.311048] env[61728]: _type = "Task" [ 801.311048] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.320423] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b572e4-36d4-6335-ca88-d9d797722619, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.567869] env[61728]: DEBUG oslo_concurrency.lockutils [req-cecd82fb-2e5c-498c-9311-67af05d93aa3 req-8456b73e-5eff-4eb7-9243-be6e7ac17d6c service nova] Releasing lock "refresh_cache-06109957-2c3b-404e-a43e-dd34ece39096" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.735018] env[61728]: DEBUG nova.compute.manager [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 801.747196] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3342366c-17c9-492e-b5d0-e974fd2a5746 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.756036] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-decad090-ec81-4202-ae3b-f7cd03050144 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.763146] env[61728]: DEBUG nova.network.neutron [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Successfully created port: 78aa30fc-a52b-4ebd-bb26-1020af74cf2e {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.798223] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1cc534-6ca3-4f1a-903d-c941ae314c0f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.807079] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08818377-9ddb-459c-b8a0-996ecf7ab464 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.828023] env[61728]: DEBUG nova.compute.provider_tree [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.832587] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b572e4-36d4-6335-ca88-d9d797722619, 'name': SearchDatastore_Task, 'duration_secs': 0.011428} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.833722] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.833722] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk. {{(pid=61728) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 801.833722] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.833933] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 801.834096] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e9799324-aebf-44e0-ae5f-4eda1358466e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.836211] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d4c034aa-e346-45ac-9a3b-7db46e3b9959 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.844416] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 801.844416] env[61728]: value = "task-464200" [ 801.844416] env[61728]: _type = "Task" [ 801.844416] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.848658] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.848856] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 801.849959] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fba35de-214d-4301-8ea1-d61203f9c6e8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.855515] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464200, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.858745] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 801.858745] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529f9354-9083-4258-d886-e5bbf4fc885e" [ 801.858745] env[61728]: _type = "Task" [ 801.858745] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.866616] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529f9354-9083-4258-d886-e5bbf4fc885e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.267029] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "f2692e22-6c14-4bb3-a080-607f0731105d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.267029] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "f2692e22-6c14-4bb3-a080-607f0731105d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.267029] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "f2692e22-6c14-4bb3-a080-607f0731105d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.334111] env[61728]: DEBUG nova.scheduler.client.report [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.354363] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464200, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467894} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.355313] env[61728]: INFO nova.virt.vmwareapi.ds_util [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk. [ 802.356133] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80468587-6ac2-40de-a52e-bcac5420dec3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.382846] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 802.384074] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1882e5b0-8891-4e3b-9290-cd542981099d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.400833] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529f9354-9083-4258-d886-e5bbf4fc885e, 'name': SearchDatastore_Task, 'duration_secs': 0.008739} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.402105] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07a4b8fe-5ec3-4a7b-ba90-c108b46d981f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.407309] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 802.407309] env[61728]: value = "task-464201" [ 802.407309] env[61728]: _type = "Task" [ 802.407309] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.407719] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 802.407719] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526b7f53-5457-055e-9425-83f2a51ffc99" [ 802.407719] env[61728]: _type = "Task" [ 802.407719] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.421012] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464201, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.422393] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526b7f53-5457-055e-9425-83f2a51ffc99, 'name': SearchDatastore_Task, 'duration_secs': 0.008646} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.422633] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.422914] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 06109957-2c3b-404e-a43e-dd34ece39096/06109957-2c3b-404e-a43e-dd34ece39096.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 802.423295] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b4efdea5-1403-44b1-a4ec-85478e7d9502 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.430559] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 802.430559] env[61728]: value = "task-464202" [ 802.430559] env[61728]: _type = "Task" [ 802.430559] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.441713] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464202, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.747307] env[61728]: DEBUG nova.compute.manager [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.789280] env[61728]: DEBUG nova.virt.hardware [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.789562] env[61728]: DEBUG nova.virt.hardware [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.793023] env[61728]: DEBUG nova.virt.hardware [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.793023] env[61728]: DEBUG nova.virt.hardware [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.793023] env[61728]: DEBUG nova.virt.hardware [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.793023] env[61728]: DEBUG nova.virt.hardware [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.793023] env[61728]: DEBUG nova.virt.hardware [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.793420] env[61728]: DEBUG nova.virt.hardware [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.793420] env[61728]: DEBUG nova.virt.hardware [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.793490] env[61728]: DEBUG nova.virt.hardware [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.793785] env[61728]: DEBUG nova.virt.hardware [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.794955] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0065f297-31c9-4bba-8a9f-f2c2477462bf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.806433] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ed3143-085f-4dfc-834b-a091a2738bf3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.839363] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.116s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.841891] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 45.183s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.846036] env[61728]: INFO nova.compute.claims [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.872438] env[61728]: INFO nova.scheduler.client.report [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Deleted allocations for instance b563e521-0e0e-45ea-bd5a-056285ea35b9 [ 802.921934] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464201, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.945066] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464202, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.427299} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.947358] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 06109957-2c3b-404e-a43e-dd34ece39096/06109957-2c3b-404e-a43e-dd34ece39096.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 802.948744] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 802.948744] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-248c2438-6e5d-4fc5-9f18-b3e49154fe46 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.955376] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 802.955376] env[61728]: value = "task-464203" [ 802.955376] env[61728]: _type = "Task" [ 802.955376] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.966713] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464203, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.325137] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.325137] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.325137] env[61728]: DEBUG nova.network.neutron [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 803.382138] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c317c9da-8110-4267-b914-c3de618c711b tempest-ServersWithSpecificFlavorTestJSON-744162176 tempest-ServersWithSpecificFlavorTestJSON-744162176-project-member] Lock "b563e521-0e0e-45ea-bd5a-056285ea35b9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 51.780s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.420463] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464201, 'name': ReconfigVM_Task, 'duration_secs': 0.531913} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.420823] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Reconfigured VM instance instance-0000003f to attach disk [datastore1] b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 803.421742] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6160204-6a5c-4e68-adb0-ffd9c5e67cd7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.449015] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42348973-dbbd-44cb-bcd3-9edbbcad5dce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.474102] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464203, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062769} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.474659] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 803.475069] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 803.475069] env[61728]: value = "task-464204" [ 803.475069] env[61728]: _type = "Task" [ 803.475069] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.475769] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df80cbc-bf77-4069-98fd-b3c91bdb7beb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.502017] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 06109957-2c3b-404e-a43e-dd34ece39096/06109957-2c3b-404e-a43e-dd34ece39096.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.505541] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a8b63cf-8638-4b06-b4d2-677c2a7e29f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.521120] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464204, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.525964] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 803.525964] env[61728]: value = "task-464205" [ 803.525964] env[61728]: _type = "Task" [ 803.525964] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.533947] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464205, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.606399] env[61728]: DEBUG nova.compute.manager [req-8312e4f1-bee5-4034-9acf-bcaab8678f6e req-05748e39-bf0f-4dde-a054-40580fe944a5 service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Received event network-vif-plugged-78aa30fc-a52b-4ebd-bb26-1020af74cf2e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 803.606700] env[61728]: DEBUG oslo_concurrency.lockutils [req-8312e4f1-bee5-4034-9acf-bcaab8678f6e req-05748e39-bf0f-4dde-a054-40580fe944a5 service nova] Acquiring lock "2976d71f-7322-4640-b734-645dbb70e0c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.606843] env[61728]: DEBUG oslo_concurrency.lockutils [req-8312e4f1-bee5-4034-9acf-bcaab8678f6e req-05748e39-bf0f-4dde-a054-40580fe944a5 service nova] Lock "2976d71f-7322-4640-b734-645dbb70e0c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.607319] env[61728]: DEBUG oslo_concurrency.lockutils [req-8312e4f1-bee5-4034-9acf-bcaab8678f6e req-05748e39-bf0f-4dde-a054-40580fe944a5 service nova] Lock "2976d71f-7322-4640-b734-645dbb70e0c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.607547] env[61728]: DEBUG nova.compute.manager [req-8312e4f1-bee5-4034-9acf-bcaab8678f6e req-05748e39-bf0f-4dde-a054-40580fe944a5 service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] No waiting events found dispatching network-vif-plugged-78aa30fc-a52b-4ebd-bb26-1020af74cf2e {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 803.607725] env[61728]: WARNING nova.compute.manager [req-8312e4f1-bee5-4034-9acf-bcaab8678f6e req-05748e39-bf0f-4dde-a054-40580fe944a5 service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Received unexpected event network-vif-plugged-78aa30fc-a52b-4ebd-bb26-1020af74cf2e for instance with vm_state building and task_state spawning. [ 803.650517] env[61728]: DEBUG nova.network.neutron [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Successfully updated port: 78aa30fc-a52b-4ebd-bb26-1020af74cf2e {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 803.992029] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464204, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.037710] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464205, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.155151] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "refresh_cache-2976d71f-7322-4640-b734-645dbb70e0c3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.155151] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "refresh_cache-2976d71f-7322-4640-b734-645dbb70e0c3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.155151] env[61728]: DEBUG nova.network.neutron [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 804.167582] env[61728]: DEBUG nova.network.neutron [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating instance_info_cache with network_info: [{"id": "cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd", "address": "fa:16:3e:fc:46:55", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc2e7ae1-f4", "ovs_interfaceid": "cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.393752] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c4b775c-feaf-4b8c-bf62-4eb225f52e91 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.403392] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16082c68-2608-4633-be3e-e2c094dbc659 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.436067] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b7aa31-984a-48fb-ad58-25298d97268d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.443866] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c2dc00-53fa-49b1-bb3f-e65753b66126 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.457078] env[61728]: DEBUG nova.compute.provider_tree [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.489534] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464204, 'name': ReconfigVM_Task, 'duration_secs': 0.713417} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.489852] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 804.489964] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-862d0eef-92a5-48f2-80c4-a0ee56344585 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.496566] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 804.496566] env[61728]: value = "task-464206" [ 804.496566] env[61728]: _type = "Task" [ 804.496566] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.504584] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464206, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.535537] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464205, 'name': ReconfigVM_Task, 'duration_secs': 0.793187} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.535844] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 06109957-2c3b-404e-a43e-dd34ece39096/06109957-2c3b-404e-a43e-dd34ece39096.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.536499] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-886675c7-1a6d-48fb-8c0a-a2796e808fa1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.543478] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 804.543478] env[61728]: value = "task-464207" [ 804.543478] env[61728]: _type = "Task" [ 804.543478] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.555746] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464207, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.673055] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.707234] env[61728]: DEBUG nova.network.neutron [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 804.871240] env[61728]: DEBUG nova.network.neutron [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Updating instance_info_cache with network_info: [{"id": "78aa30fc-a52b-4ebd-bb26-1020af74cf2e", "address": "fa:16:3e:16:c0:48", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78aa30fc-a5", "ovs_interfaceid": "78aa30fc-a52b-4ebd-bb26-1020af74cf2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.960724] env[61728]: DEBUG nova.scheduler.client.report [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.008729] env[61728]: DEBUG oslo_vmware.api [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464206, 'name': PowerOnVM_Task, 'duration_secs': 0.432649} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.009051] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 805.013661] env[61728]: DEBUG nova.compute.manager [None req-e8b131ea-9066-4ea8-9f9e-75b8eb9be559 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.015102] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845af35a-9995-460e-8cec-65f30bd41ba4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.055748] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464207, 'name': Rename_Task, 'duration_secs': 0.155221} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.055748] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 805.055748] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b3a91447-cf71-4585-b6e9-221b11cd3d05 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.063868] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 805.063868] env[61728]: value = "task-464208" [ 805.063868] env[61728]: _type = "Task" [ 805.063868] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.077420] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464208, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.202604] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9862a86a-8804-4a3e-a707-607525358da2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.222862] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c27501ef-5531-44ad-a741-d54e4cb463f7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.229880] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating instance 'f2692e22-6c14-4bb3-a080-607f0731105d' progress to 83 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 805.375081] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "refresh_cache-2976d71f-7322-4640-b734-645dbb70e0c3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.375446] env[61728]: DEBUG nova.compute.manager [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Instance network_info: |[{"id": "78aa30fc-a52b-4ebd-bb26-1020af74cf2e", "address": "fa:16:3e:16:c0:48", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78aa30fc-a5", "ovs_interfaceid": "78aa30fc-a52b-4ebd-bb26-1020af74cf2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 805.375905] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:c0:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd733acc2-07d0-479e-918c-ec8a21925389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '78aa30fc-a52b-4ebd-bb26-1020af74cf2e', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.383854] env[61728]: DEBUG oslo.service.loopingcall [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.384073] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 805.384497] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9db9e2de-5b29-4c5b-81e6-176c315c58f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.405123] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.405123] env[61728]: value = "task-464209" [ 805.405123] env[61728]: _type = "Task" [ 805.405123] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.413405] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464209, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.467552] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.626s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.471179] env[61728]: DEBUG nova.compute.manager [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 805.471179] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 46.356s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.573730] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464208, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.736883] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 805.737263] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-33e9d98a-f89b-469b-9ee3-686a743a3c01 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.745684] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 805.745684] env[61728]: value = "task-464210" [ 805.745684] env[61728]: _type = "Task" [ 805.745684] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.753592] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464210, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.917912] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464209, 'name': CreateVM_Task, 'duration_secs': 0.336332} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.918353] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 805.919103] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.919406] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.919643] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 805.919956] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e69e54da-9a61-4991-b7d8-ce6b82182f0f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.924735] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 805.924735] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f50c48-675d-99d4-1d3a-15fef1fd92fb" [ 805.924735] env[61728]: _type = "Task" [ 805.924735] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.933269] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f50c48-675d-99d4-1d3a-15fef1fd92fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.978225] env[61728]: DEBUG nova.compute.utils [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 805.979822] env[61728]: DEBUG nova.objects.instance [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lazy-loading 'migration_context' on Instance uuid 1d9114cb-a0af-4874-962f-27237b3c89cc {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 805.985320] env[61728]: DEBUG nova.compute.manager [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Not allocating networking since 'none' was specified. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 806.077022] env[61728]: DEBUG oslo_vmware.api [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464208, 'name': PowerOnVM_Task, 'duration_secs': 0.731657} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.077660] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 806.078026] env[61728]: INFO nova.compute.manager [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Took 8.26 seconds to spawn the instance on the hypervisor. [ 806.078321] env[61728]: DEBUG nova.compute.manager [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 806.079270] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63aa7e55-5e1a-4825-adaf-219b721c5d55 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.180189] env[61728]: DEBUG nova.compute.manager [req-cc6629db-099f-465f-8c6e-a0a84e6589f2 req-dd85f83a-5a59-45b3-b3de-31aede4ac478 service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Received event network-changed-78aa30fc-a52b-4ebd-bb26-1020af74cf2e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 806.180189] env[61728]: DEBUG nova.compute.manager [req-cc6629db-099f-465f-8c6e-a0a84e6589f2 req-dd85f83a-5a59-45b3-b3de-31aede4ac478 service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Refreshing instance network info cache due to event network-changed-78aa30fc-a52b-4ebd-bb26-1020af74cf2e. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 806.180189] env[61728]: DEBUG oslo_concurrency.lockutils [req-cc6629db-099f-465f-8c6e-a0a84e6589f2 req-dd85f83a-5a59-45b3-b3de-31aede4ac478 service nova] Acquiring lock "refresh_cache-2976d71f-7322-4640-b734-645dbb70e0c3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.180189] env[61728]: DEBUG oslo_concurrency.lockutils [req-cc6629db-099f-465f-8c6e-a0a84e6589f2 req-dd85f83a-5a59-45b3-b3de-31aede4ac478 service nova] Acquired lock "refresh_cache-2976d71f-7322-4640-b734-645dbb70e0c3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.180189] env[61728]: DEBUG nova.network.neutron [req-cc6629db-099f-465f-8c6e-a0a84e6589f2 req-dd85f83a-5a59-45b3-b3de-31aede4ac478 service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Refreshing network info cache for port 78aa30fc-a52b-4ebd-bb26-1020af74cf2e {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 806.259396] env[61728]: DEBUG oslo_vmware.api [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464210, 'name': PowerOnVM_Task, 'duration_secs': 0.449968} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.259872] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 806.260330] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a406b293-c4dd-4225-ac2f-da510c5de23d tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating instance 'f2692e22-6c14-4bb3-a080-607f0731105d' progress to 100 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 806.440783] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f50c48-675d-99d4-1d3a-15fef1fd92fb, 'name': SearchDatastore_Task, 'duration_secs': 0.020252} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.440783] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.440783] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.440783] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.440783] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.440783] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.440783] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc0bb6ab-2894-43a3-b439-855dfb1bceff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.456021] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.456021] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 806.456021] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3920a83d-1b9f-41df-b125-f2c8838bfea3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.466659] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 806.466659] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524bde90-b71a-296d-9e92-cd913a314179" [ 806.466659] env[61728]: _type = "Task" [ 806.466659] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.474233] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524bde90-b71a-296d-9e92-cd913a314179, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.488789] env[61728]: DEBUG nova.compute.manager [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 806.600235] env[61728]: INFO nova.compute.manager [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Took 59.02 seconds to build instance. [ 806.948210] env[61728]: DEBUG nova.compute.manager [req-4528e30b-5369-461c-a039-604db6edc232 req-46c6f8d2-bac3-48a5-b24a-ab1e0a5b2da8 service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Received event network-changed-920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 806.948437] env[61728]: DEBUG nova.compute.manager [req-4528e30b-5369-461c-a039-604db6edc232 req-46c6f8d2-bac3-48a5-b24a-ab1e0a5b2da8 service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Refreshing instance network info cache due to event network-changed-920c9509-d0d8-464d-a5e2-bef1ba4f5637. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 806.948680] env[61728]: DEBUG oslo_concurrency.lockutils [req-4528e30b-5369-461c-a039-604db6edc232 req-46c6f8d2-bac3-48a5-b24a-ab1e0a5b2da8 service nova] Acquiring lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.948853] env[61728]: DEBUG oslo_concurrency.lockutils [req-4528e30b-5369-461c-a039-604db6edc232 req-46c6f8d2-bac3-48a5-b24a-ab1e0a5b2da8 service nova] Acquired lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.949055] env[61728]: DEBUG nova.network.neutron [req-4528e30b-5369-461c-a039-604db6edc232 req-46c6f8d2-bac3-48a5-b24a-ab1e0a5b2da8 service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Refreshing network info cache for port 920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 806.980833] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524bde90-b71a-296d-9e92-cd913a314179, 'name': SearchDatastore_Task, 'duration_secs': 0.033133} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.981942] env[61728]: DEBUG nova.network.neutron [req-cc6629db-099f-465f-8c6e-a0a84e6589f2 req-dd85f83a-5a59-45b3-b3de-31aede4ac478 service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Updated VIF entry in instance network info cache for port 78aa30fc-a52b-4ebd-bb26-1020af74cf2e. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 806.982304] env[61728]: DEBUG nova.network.neutron [req-cc6629db-099f-465f-8c6e-a0a84e6589f2 req-dd85f83a-5a59-45b3-b3de-31aede4ac478 service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Updating instance_info_cache with network_info: [{"id": "78aa30fc-a52b-4ebd-bb26-1020af74cf2e", "address": "fa:16:3e:16:c0:48", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78aa30fc-a5", "ovs_interfaceid": "78aa30fc-a52b-4ebd-bb26-1020af74cf2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.986417] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e39ea28a-4096-4015-9cdb-195522d41188 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.007373] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 807.007373] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528b121b-e848-252a-d8ac-965d6320a683" [ 807.007373] env[61728]: _type = "Task" [ 807.007373] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.022231] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528b121b-e848-252a-d8ac-965d6320a683, 'name': SearchDatastore_Task, 'duration_secs': 0.010607} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.022231] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.022231] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2976d71f-7322-4640-b734-645dbb70e0c3/2976d71f-7322-4640-b734-645dbb70e0c3.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 807.022231] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b06291ba-25af-43ab-956d-630d98593524 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.027798] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 807.027798] env[61728]: value = "task-464211" [ 807.027798] env[61728]: _type = "Task" [ 807.027798] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.047795] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464211, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.103574] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1fd6985f-f754-434a-a15f-a44f9589cba4 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "06109957-2c3b-404e-a43e-dd34ece39096" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.508s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.124465] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c59ac358-429f-4f05-ad4c-4ecd9642bbd9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.135577] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc05be85-bb9e-486a-b372-2fbe14d67cb0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.168385] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c39da71-f893-456b-be62-f7918ad30b46 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.176865] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77792cc-7f3c-4d1c-8585-0db25f6000d1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.191272] env[61728]: DEBUG nova.compute.provider_tree [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.485368] env[61728]: DEBUG oslo_concurrency.lockutils [req-cc6629db-099f-465f-8c6e-a0a84e6589f2 req-dd85f83a-5a59-45b3-b3de-31aede4ac478 service nova] Releasing lock "refresh_cache-2976d71f-7322-4640-b734-645dbb70e0c3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.503069] env[61728]: DEBUG nova.compute.manager [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 807.538525] env[61728]: DEBUG nova.virt.hardware [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 807.538854] env[61728]: DEBUG nova.virt.hardware [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 807.539232] env[61728]: DEBUG nova.virt.hardware [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 807.539534] env[61728]: DEBUG nova.virt.hardware [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 807.539717] env[61728]: DEBUG nova.virt.hardware [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 807.541035] env[61728]: DEBUG nova.virt.hardware [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 807.541035] env[61728]: DEBUG nova.virt.hardware [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 807.541035] env[61728]: DEBUG nova.virt.hardware [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 807.541035] env[61728]: DEBUG nova.virt.hardware [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 807.541035] env[61728]: DEBUG nova.virt.hardware [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 807.541035] env[61728]: DEBUG nova.virt.hardware [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 807.545924] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec60f5fb-9c33-4c9f-9f67-b31009f408c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.548691] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464211, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505052} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.549181] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2976d71f-7322-4640-b734-645dbb70e0c3/2976d71f-7322-4640-b734-645dbb70e0c3.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 807.549286] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.550040] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5761518-6411-4770-8239-dc99cf3dafe9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.556491] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac949d8-d65c-4281-90ed-dabb041bd6a1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.562334] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 807.562334] env[61728]: value = "task-464212" [ 807.562334] env[61728]: _type = "Task" [ 807.562334] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.576455] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Instance VIF info [] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.582598] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Creating folder: Project (471fd37514e84e96ad10e53dafec2ba8). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 807.585918] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e5d15665-4073-4fb0-a4b5-a91564b8d81d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.591599] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464212, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.600088] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Created folder: Project (471fd37514e84e96ad10e53dafec2ba8) in parent group-v121913. [ 807.600356] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Creating folder: Instances. Parent ref: group-v122107. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 807.600638] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-54427156-4430-4c0a-b20e-3179bd9968ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.609756] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Created folder: Instances in parent group-v122107. [ 807.610087] env[61728]: DEBUG oslo.service.loopingcall [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.610313] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 807.610567] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-040cf04a-1a3f-4129-955c-c014d9aa792b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.633845] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.633845] env[61728]: value = "task-464215" [ 807.633845] env[61728]: _type = "Task" [ 807.633845] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.641540] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464215, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.696053] env[61728]: DEBUG nova.scheduler.client.report [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 807.854478] env[61728]: DEBUG nova.network.neutron [req-4528e30b-5369-461c-a039-604db6edc232 req-46c6f8d2-bac3-48a5-b24a-ab1e0a5b2da8 service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Updated VIF entry in instance network info cache for port 920c9509-d0d8-464d-a5e2-bef1ba4f5637. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 807.854567] env[61728]: DEBUG nova.network.neutron [req-4528e30b-5369-461c-a039-604db6edc232 req-46c6f8d2-bac3-48a5-b24a-ab1e0a5b2da8 service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Updating instance_info_cache with network_info: [{"id": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "address": "fa:16:3e:c7:4f:2f", "network": {"id": "15423738-1cd1-4170-9505-8cd6b2a672ca", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1314838493-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f742e5f95a3f4ef8b407894687866958", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76e60ff4-204c-4f48-bd0e-2d5fa0a812ef", "external-id": "nsx-vlan-transportzone-854", "segmentation_id": 854, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920c9509-d0", "ovs_interfaceid": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.082078] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464212, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069369} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.082391] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.083273] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af6df5fa-8343-4c9b-94e8-f87dfd96b3b1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.113375] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 2976d71f-7322-4640-b734-645dbb70e0c3/2976d71f-7322-4640-b734-645dbb70e0c3.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.113810] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5703e382-a608-4362-b11b-838c24e37ec1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.140524] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 808.140524] env[61728]: value = "task-464216" [ 808.140524] env[61728]: _type = "Task" [ 808.140524] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.149497] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464215, 'name': CreateVM_Task, 'duration_secs': 0.298701} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.150089] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 808.150764] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.151066] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.151353] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 808.155183] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10d56df4-1ca7-4aaf-a309-f0d32c79e1c0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.157251] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464216, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.161302] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 808.161302] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d2fa88-ed6e-aac1-64a4-94c141b17811" [ 808.161302] env[61728]: _type = "Task" [ 808.161302] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.170411] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d2fa88-ed6e-aac1-64a4-94c141b17811, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.357139] env[61728]: DEBUG oslo_concurrency.lockutils [req-4528e30b-5369-461c-a039-604db6edc232 req-46c6f8d2-bac3-48a5-b24a-ab1e0a5b2da8 service nova] Releasing lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.652911] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464216, 'name': ReconfigVM_Task, 'duration_secs': 0.369771} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.653265] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 2976d71f-7322-4640-b734-645dbb70e0c3/2976d71f-7322-4640-b734-645dbb70e0c3.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 808.654132] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e1bb5f2f-9541-4e69-856d-6d66bd758cb7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.664320] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 808.664320] env[61728]: value = "task-464217" [ 808.664320] env[61728]: _type = "Task" [ 808.664320] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.683170] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464217, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.683170] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d2fa88-ed6e-aac1-64a4-94c141b17811, 'name': SearchDatastore_Task, 'duration_secs': 0.018874} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.683416] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.683661] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 808.683903] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.684086] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.684277] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 808.684545] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9de8042e-e0e3-4167-90f5-c93216c81bc9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.694370] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 808.694688] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 808.695410] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58e75034-7865-4985-8e5e-dfaa73a8dae5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.704113] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 808.704113] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521aff9e-4edd-7dd1-adbe-5d96abbd05eb" [ 808.704113] env[61728]: _type = "Task" [ 808.704113] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.710290] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.239s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.716819] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 45.719s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.717230] env[61728]: DEBUG nova.objects.instance [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: 2e83145a-2481-411d-b037-a305a209074e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61728) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 808.730958] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521aff9e-4edd-7dd1-adbe-5d96abbd05eb, 'name': SearchDatastore_Task, 'duration_secs': 0.018869} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.732148] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cebc0e3-b7fa-428b-b687-26c7d625a0d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.740914] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 808.740914] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5233d8da-6eec-8e48-2062-d1d8a86da3d3" [ 808.740914] env[61728]: _type = "Task" [ 808.740914] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.753540] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5233d8da-6eec-8e48-2062-d1d8a86da3d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.764175] env[61728]: DEBUG nova.compute.manager [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 808.764175] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccdc5034-d8bb-412a-b44c-680adb8ecb78 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.075365] env[61728]: DEBUG oslo_concurrency.lockutils [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "f2692e22-6c14-4bb3-a080-607f0731105d" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.075636] env[61728]: DEBUG oslo_concurrency.lockutils [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "f2692e22-6c14-4bb3-a080-607f0731105d" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.075835] env[61728]: DEBUG nova.compute.manager [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Going to confirm migration 4 {{(pid=61728) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 809.180662] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464217, 'name': Rename_Task, 'duration_secs': 0.248653} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.181784] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 809.182444] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4726684-9623-42f9-a6fa-ae4bf4b689e8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.189148] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 809.189148] env[61728]: value = "task-464218" [ 809.189148] env[61728]: _type = "Task" [ 809.189148] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.198992] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464218, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.254020] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5233d8da-6eec-8e48-2062-d1d8a86da3d3, 'name': SearchDatastore_Task, 'duration_secs': 0.016833} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.254020] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.254020] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 6d55bb01-7855-4d86-ac57-ecae15ea590f/6d55bb01-7855-4d86-ac57-ecae15ea590f.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 809.254020] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b41fa69a-f629-4682-bb6d-431c99e7311e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.260931] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 809.260931] env[61728]: value = "task-464219" [ 809.260931] env[61728]: _type = "Task" [ 809.260931] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.271272] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464219, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.276720] env[61728]: INFO nova.compute.manager [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] instance snapshotting [ 809.281200] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac92651-565e-487a-8072-3bea7e24763d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.302695] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baaefc20-f62d-496f-807d-516f51464adb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.642381] env[61728]: DEBUG oslo_concurrency.lockutils [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.642539] env[61728]: DEBUG oslo_concurrency.lockutils [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.642717] env[61728]: DEBUG nova.network.neutron [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 809.642905] env[61728]: DEBUG nova.objects.instance [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lazy-loading 'info_cache' on Instance uuid f2692e22-6c14-4bb3-a080-607f0731105d {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 809.696390] env[61728]: DEBUG nova.compute.manager [req-0305aaae-7ee3-424c-8894-b812ba654b6d req-d4666cbd-6d7c-46a1-a7a9-74014537a936 service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Received event network-changed-920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 809.696390] env[61728]: DEBUG nova.compute.manager [req-0305aaae-7ee3-424c-8894-b812ba654b6d req-d4666cbd-6d7c-46a1-a7a9-74014537a936 service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Refreshing instance network info cache due to event network-changed-920c9509-d0d8-464d-a5e2-bef1ba4f5637. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 809.696390] env[61728]: DEBUG oslo_concurrency.lockutils [req-0305aaae-7ee3-424c-8894-b812ba654b6d req-d4666cbd-6d7c-46a1-a7a9-74014537a936 service nova] Acquiring lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.696507] env[61728]: DEBUG oslo_concurrency.lockutils [req-0305aaae-7ee3-424c-8894-b812ba654b6d req-d4666cbd-6d7c-46a1-a7a9-74014537a936 service nova] Acquired lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.696566] env[61728]: DEBUG nova.network.neutron [req-0305aaae-7ee3-424c-8894-b812ba654b6d req-d4666cbd-6d7c-46a1-a7a9-74014537a936 service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Refreshing network info cache for port 920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 809.706845] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464218, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.728363] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fdb0603e-f457-46c5-ae3d-47bccfe13a57 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.729837] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 42.469s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.730061] env[61728]: DEBUG nova.objects.instance [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lazy-loading 'resources' on Instance uuid 2d389d5e-a53f-424e-90af-5ed05e2f93c7 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 809.773781] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464219, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.816041] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 809.816041] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9569a2b6-8a6a-4720-ab7c-b2338b8c08ac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.823117] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 809.823117] env[61728]: value = "task-464220" [ 809.823117] env[61728]: _type = "Task" [ 809.823117] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.834708] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464220, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.202579] env[61728]: DEBUG oslo_vmware.api [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464218, 'name': PowerOnVM_Task, 'duration_secs': 0.889474} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.202887] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 810.202986] env[61728]: INFO nova.compute.manager [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Took 7.46 seconds to spawn the instance on the hypervisor. [ 810.203145] env[61728]: DEBUG nova.compute.manager [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.208021] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72ddb8ce-d7c1-4d95-9e15-698c21763118 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.270868] env[61728]: INFO nova.compute.manager [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Swapping old allocation on dict_keys(['e7ceb92f-072b-409e-b888-6fe0676b32f1']) held by migration 9939d541-20e8-411b-981f-657861975b29 for instance [ 810.281833] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464219, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.91611} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.282132] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 6d55bb01-7855-4d86-ac57-ecae15ea590f/6d55bb01-7855-4d86-ac57-ecae15ea590f.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 810.282353] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 810.282651] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c71d425b-66be-43ff-afca-00846edc8b39 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.295020] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 810.295020] env[61728]: value = "task-464221" [ 810.295020] env[61728]: _type = "Task" [ 810.295020] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.303632] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464221, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.304694] env[61728]: DEBUG nova.scheduler.client.report [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Overwriting current allocation {'allocations': {'e7ceb92f-072b-409e-b888-6fe0676b32f1': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 103}}, 'project_id': '2ac86255b8224fc095ca68631066afbe', 'user_id': '244030ae678b4d8588086675b6d23718', 'consumer_generation': 1} on consumer 1d9114cb-a0af-4874-962f-27237b3c89cc {{(pid=61728) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 810.334322] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464220, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.408655] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.408850] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquired lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.409049] env[61728]: DEBUG nova.network.neutron [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 810.466760] env[61728]: DEBUG nova.network.neutron [req-0305aaae-7ee3-424c-8894-b812ba654b6d req-d4666cbd-6d7c-46a1-a7a9-74014537a936 service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Updated VIF entry in instance network info cache for port 920c9509-d0d8-464d-a5e2-bef1ba4f5637. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 810.466760] env[61728]: DEBUG nova.network.neutron [req-0305aaae-7ee3-424c-8894-b812ba654b6d req-d4666cbd-6d7c-46a1-a7a9-74014537a936 service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Updating instance_info_cache with network_info: [{"id": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "address": "fa:16:3e:c7:4f:2f", "network": {"id": "15423738-1cd1-4170-9505-8cd6b2a672ca", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1314838493-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.236", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f742e5f95a3f4ef8b407894687866958", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76e60ff4-204c-4f48-bd0e-2d5fa0a812ef", "external-id": "nsx-vlan-transportzone-854", "segmentation_id": 854, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920c9509-d0", "ovs_interfaceid": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.730564] env[61728]: INFO nova.compute.manager [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Took 56.91 seconds to build instance. [ 810.805763] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464221, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068063} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.806100] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 810.806887] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03525773-1673-4910-8349-f1801f8ae25a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.827402] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 6d55bb01-7855-4d86-ac57-ecae15ea590f/6d55bb01-7855-4d86-ac57-ecae15ea590f.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 810.828609] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9d0b8f1-d000-44bc-8bfb-fb3796ce32cb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.843203] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3205a5de-99c8-4350-a51b-6a247d534416 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.860492] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d64e4ca9-d82c-4d41-90f3-5ed366bcff84 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.863848] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464220, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.864201] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 810.864201] env[61728]: value = "task-464222" [ 810.864201] env[61728]: _type = "Task" [ 810.864201] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.896215] env[61728]: DEBUG nova.network.neutron [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating instance_info_cache with network_info: [{"id": "cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd", "address": "fa:16:3e:fc:46:55", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc2e7ae1-f4", "ovs_interfaceid": "cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.898464] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe17730-2b8b-4377-88c2-94d50b72ce1a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.904633] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464222, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.910449] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab22850-b31f-4f46-aa4d-3ce3ce3a5b23 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.927646] env[61728]: DEBUG nova.compute.provider_tree [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 810.968574] env[61728]: DEBUG oslo_concurrency.lockutils [req-0305aaae-7ee3-424c-8894-b812ba654b6d req-d4666cbd-6d7c-46a1-a7a9-74014537a936 service nova] Releasing lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.136456] env[61728]: DEBUG nova.network.neutron [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance_info_cache with network_info: [{"id": "4d70757b-d260-4aa3-8914-a38eb567f7b8", "address": "fa:16:3e:2a:f5:67", "network": {"id": "ae586e2e-1b64-4d5e-8598-2ea3c22ffddc", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.112", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fb11ba9be5014418bbf48b9cc32669bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "457c42cd-4ddb-4374-923e-d419b7f6eaff", "external-id": "nsx-vlan-transportzone-575", "segmentation_id": 575, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d70757b-d2", "ovs_interfaceid": "4d70757b-d260-4aa3-8914-a38eb567f7b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.232272] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a594570c-357e-4133-a316-a4ebf2405d07 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "2976d71f-7322-4640-b734-645dbb70e0c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.571s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.339479] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464220, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.374044] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464222, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.402166] env[61728]: DEBUG oslo_concurrency.lockutils [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "refresh_cache-f2692e22-6c14-4bb3-a080-607f0731105d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.402840] env[61728]: DEBUG nova.objects.instance [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lazy-loading 'migration_context' on Instance uuid f2692e22-6c14-4bb3-a080-607f0731105d {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 811.451921] env[61728]: ERROR nova.scheduler.client.report [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [req-af1a084a-8f57-4ec7-a742-7306c19bede4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 112, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-af1a084a-8f57-4ec7-a742-7306c19bede4"}]} [ 811.471535] env[61728]: DEBUG nova.scheduler.client.report [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 811.495097] env[61728]: DEBUG nova.scheduler.client.report [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 811.495347] env[61728]: DEBUG nova.compute.provider_tree [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 811.510922] env[61728]: DEBUG nova.scheduler.client.report [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 811.532358] env[61728]: DEBUG nova.scheduler.client.report [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 811.643632] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Releasing lock "refresh_cache-1d9114cb-a0af-4874-962f-27237b3c89cc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.644128] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 811.645370] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-931550c2-0e7c-4fa3-b463-4487cd49d3ce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.652474] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 811.652474] env[61728]: value = "task-464223" [ 811.652474] env[61728]: _type = "Task" [ 811.652474] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.665532] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464223, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.841452] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464220, 'name': CreateSnapshot_Task, 'duration_secs': 1.554102} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.844542] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 811.848204] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908413e2-ceb8-48b5-ba40-c9763730b562 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.874124] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464222, 'name': ReconfigVM_Task, 'duration_secs': 0.90456} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.874581] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 6d55bb01-7855-4d86-ac57-ecae15ea590f/6d55bb01-7855-4d86-ac57-ecae15ea590f.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 811.875615] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-37b2b5ba-53a5-4fa5-b068-bf4c8a01a08c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.884344] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 811.884344] env[61728]: value = "task-464224" [ 811.884344] env[61728]: _type = "Task" [ 811.884344] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.899851] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464224, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.906933] env[61728]: DEBUG nova.objects.base [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 811.908842] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c54f1a6-fbae-432c-bcf6-a764f9975f81 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.935909] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8581e1aa-3b1d-4a45-8587-9f1d25934495 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.941716] env[61728]: DEBUG oslo_vmware.api [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 811.941716] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5269cf2d-826d-3ce0-3e28-5ddb3d83cb4f" [ 811.941716] env[61728]: _type = "Task" [ 811.941716] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.951555] env[61728]: DEBUG oslo_vmware.api [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5269cf2d-826d-3ce0-3e28-5ddb3d83cb4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.149220] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624e86e0-b09f-43ac-96d5-de2d74df8c22 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.159633] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7986f98-1965-43f5-af5f-2a0e83bfdd81 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.165930] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464223, 'name': PowerOffVM_Task, 'duration_secs': 0.356249} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.166641] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 812.167341] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:24:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='ec856338-a71e-43d6-aa12-96f76d31f40f',id=36,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1304387059',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 812.167620] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 812.167722] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 812.167934] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 812.168139] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 812.168301] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 812.168516] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 812.168679] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 812.168909] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 812.169158] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 812.169262] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 812.198150] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-79190916-9937-4aa7-a7d9-b17069aeec2f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.209396] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65e03af-5bb1-4517-9217-b11117ce91fc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.218269] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378f0e67-cfe3-4722-97fc-e8d52396954d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.223068] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 812.223068] env[61728]: value = "task-464225" [ 812.223068] env[61728]: _type = "Task" [ 812.223068] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.235052] env[61728]: DEBUG nova.compute.provider_tree [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.241901] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464225, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.372114] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 812.372114] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b3c0e6ab-bbe1-4020-a975-9d4f55351563 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.379017] env[61728]: DEBUG nova.compute.manager [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Received event network-changed-920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 812.379417] env[61728]: DEBUG nova.compute.manager [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Refreshing instance network info cache due to event network-changed-920c9509-d0d8-464d-a5e2-bef1ba4f5637. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 812.379753] env[61728]: DEBUG oslo_concurrency.lockutils [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] Acquiring lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.380013] env[61728]: DEBUG oslo_concurrency.lockutils [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] Acquired lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.380291] env[61728]: DEBUG nova.network.neutron [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Refreshing network info cache for port 920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 812.383282] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 812.383282] env[61728]: value = "task-464226" [ 812.383282] env[61728]: _type = "Task" [ 812.383282] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.396597] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464226, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.400201] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464224, 'name': Rename_Task, 'duration_secs': 0.268195} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.400466] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 812.401136] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ccea24b0-0844-40b2-861a-5a08ed70c4c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.407599] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 812.407599] env[61728]: value = "task-464227" [ 812.407599] env[61728]: _type = "Task" [ 812.407599] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.416159] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464227, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.452356] env[61728]: DEBUG oslo_vmware.api [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5269cf2d-826d-3ce0-3e28-5ddb3d83cb4f, 'name': SearchDatastore_Task, 'duration_secs': 0.009114} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.452678] env[61728]: DEBUG oslo_concurrency.lockutils [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.732928] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464225, 'name': ReconfigVM_Task, 'duration_secs': 0.298815} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.734034] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b4c764-8e4f-47db-8cdc-6b0409313e21 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.753282] env[61728]: DEBUG nova.scheduler.client.report [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.757017] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:24:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='ec856338-a71e-43d6-aa12-96f76d31f40f',id=36,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1304387059',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 812.757306] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 812.757481] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 812.757722] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 812.757915] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 812.758126] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 812.758382] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 812.758583] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 812.758793] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 812.759501] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 812.759765] env[61728]: DEBUG nova.virt.hardware [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 812.761019] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46013a96-cf24-418d-9128-000f1aa0df0e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.766990] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 812.766990] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5213eced-3d40-5c33-ebf2-1c848ac32c8d" [ 812.766990] env[61728]: _type = "Task" [ 812.766990] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.775967] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5213eced-3d40-5c33-ebf2-1c848ac32c8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.901669] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464226, 'name': CloneVM_Task} progress is 93%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.919813] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464227, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.108949] env[61728]: DEBUG nova.network.neutron [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Updated VIF entry in instance network info cache for port 920c9509-d0d8-464d-a5e2-bef1ba4f5637. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 813.109318] env[61728]: DEBUG nova.network.neutron [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Updating instance_info_cache with network_info: [{"id": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "address": "fa:16:3e:c7:4f:2f", "network": {"id": "15423738-1cd1-4170-9505-8cd6b2a672ca", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1314838493-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f742e5f95a3f4ef8b407894687866958", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76e60ff4-204c-4f48-bd0e-2d5fa0a812ef", "external-id": "nsx-vlan-transportzone-854", "segmentation_id": 854, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920c9509-d0", "ovs_interfaceid": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.262492] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.533s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.265605] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 45.898s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.265926] env[61728]: DEBUG nova.objects.instance [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lazy-loading 'resources' on Instance uuid 2e83145a-2481-411d-b037-a305a209074e {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 813.277330] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5213eced-3d40-5c33-ebf2-1c848ac32c8d, 'name': SearchDatastore_Task, 'duration_secs': 0.019031} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.284184] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Reconfiguring VM instance instance-0000002b to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 813.285063] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da8cc768-6206-49e2-b87c-86cf6a77d8f9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.300058] env[61728]: INFO nova.scheduler.client.report [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleted allocations for instance 2d389d5e-a53f-424e-90af-5ed05e2f93c7 [ 813.306835] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 813.306835] env[61728]: value = "task-464228" [ 813.306835] env[61728]: _type = "Task" [ 813.306835] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.315635] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464228, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.401796] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464226, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.419332] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464227, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.581179] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "2976d71f-7322-4640-b734-645dbb70e0c3" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.581560] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "2976d71f-7322-4640-b734-645dbb70e0c3" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.581900] env[61728]: INFO nova.compute.manager [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Shelving [ 813.612551] env[61728]: DEBUG oslo_concurrency.lockutils [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] Releasing lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.613554] env[61728]: DEBUG nova.compute.manager [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Received event network-changed-920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 813.613554] env[61728]: DEBUG nova.compute.manager [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Refreshing instance network info cache due to event network-changed-920c9509-d0d8-464d-a5e2-bef1ba4f5637. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 813.613554] env[61728]: DEBUG oslo_concurrency.lockutils [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] Acquiring lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.613964] env[61728]: DEBUG oslo_concurrency.lockutils [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] Acquired lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.613964] env[61728]: DEBUG nova.network.neutron [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Refreshing network info cache for port 920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 813.785329] env[61728]: DEBUG nova.objects.instance [None req-3590b696-2446-4ab2-9a45-b9d95d887cf5 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Lazy-loading 'flavor' on Instance uuid e3116a87-ce12-4ec8-b6fa-70ec2127ebb0 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 813.811703] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90dd72a9-5af4-4c5b-aaa5-28fefd1a3b5d tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "2d389d5e-a53f-424e-90af-5ed05e2f93c7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 49.470s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.821722] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464228, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.902416] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464226, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.919945] env[61728]: DEBUG oslo_vmware.api [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464227, 'name': PowerOnVM_Task, 'duration_secs': 1.266042} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.920417] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 813.920677] env[61728]: INFO nova.compute.manager [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Took 6.42 seconds to spawn the instance on the hypervisor. [ 813.920892] env[61728]: DEBUG nova.compute.manager [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 813.921736] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc3d27b-52a8-43ff-bb08-0b680ff7898e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.091917] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 814.092276] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb4c61d6-8ebb-4fde-88e1-ec195e54cc4c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.102686] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 814.102686] env[61728]: value = "task-464229" [ 814.102686] env[61728]: _type = "Task" [ 814.102686] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.114422] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464229, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.252772] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquiring lock "b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.253049] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Lock "b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.253278] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquiring lock "b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.253490] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Lock "b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.253637] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Lock "b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.256658] env[61728]: INFO nova.compute.manager [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Terminating instance [ 814.260205] env[61728]: DEBUG nova.compute.manager [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 814.260205] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 814.260660] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc65de7-2305-4896-a918-50c47c838fe9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.271397] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 814.271999] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-667aff13-7e8d-4e6f-9e3a-c30e6b4ad841 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.279135] env[61728]: DEBUG oslo_vmware.api [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 814.279135] env[61728]: value = "task-464230" [ 814.279135] env[61728]: _type = "Task" [ 814.279135] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.280397] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2768da2c-a98b-4314-ae9f-e1dbbd5dd4b4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.290261] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3590b696-2446-4ab2-9a45-b9d95d887cf5 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquiring lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.290476] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3590b696-2446-4ab2-9a45-b9d95d887cf5 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquired lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.300454] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42e92fd-7a2a-4dee-a695-6e6728ca87cf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.304403] env[61728]: DEBUG oslo_vmware.api [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464230, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.338804] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7799fd12-7899-43d3-8ab2-c0eca2e96df2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.357949] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464228, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.362841] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c127d3c-b8ac-46c8-8a66-91a94a65fc05 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.379667] env[61728]: DEBUG nova.compute.provider_tree [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.398613] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464226, 'name': CloneVM_Task, 'duration_secs': 1.994084} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.398919] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Created linked-clone VM from snapshot [ 814.399744] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd076474-35f8-4089-9285-c0aa78f3bb63 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.412309] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Uploading image 82a9852c-0eb3-4602-b891-4c6c26b90c26 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 814.442863] env[61728]: INFO nova.compute.manager [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Took 56.80 seconds to build instance. [ 814.450984] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 814.450984] env[61728]: value = "vm-122111" [ 814.450984] env[61728]: _type = "VirtualMachine" [ 814.450984] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 814.453196] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-af54bafa-e038-42a8-9aa8-34db330d99d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.456960] env[61728]: DEBUG nova.network.neutron [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Updated VIF entry in instance network info cache for port 920c9509-d0d8-464d-a5e2-bef1ba4f5637. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 814.457624] env[61728]: DEBUG nova.network.neutron [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Updating instance_info_cache with network_info: [{"id": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "address": "fa:16:3e:c7:4f:2f", "network": {"id": "15423738-1cd1-4170-9505-8cd6b2a672ca", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1314838493-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f742e5f95a3f4ef8b407894687866958", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76e60ff4-204c-4f48-bd0e-2d5fa0a812ef", "external-id": "nsx-vlan-transportzone-854", "segmentation_id": 854, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920c9509-d0", "ovs_interfaceid": "920c9509-d0d8-464d-a5e2-bef1ba4f5637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.460338] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lease: (returnval){ [ 814.460338] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5296edfe-114b-8cda-26e6-2b2cb837079c" [ 814.460338] env[61728]: _type = "HttpNfcLease" [ 814.460338] env[61728]: } obtained for exporting VM: (result){ [ 814.460338] env[61728]: value = "vm-122111" [ 814.460338] env[61728]: _type = "VirtualMachine" [ 814.460338] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 814.460593] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the lease: (returnval){ [ 814.460593] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5296edfe-114b-8cda-26e6-2b2cb837079c" [ 814.460593] env[61728]: _type = "HttpNfcLease" [ 814.460593] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 814.470034] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 814.470034] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5296edfe-114b-8cda-26e6-2b2cb837079c" [ 814.470034] env[61728]: _type = "HttpNfcLease" [ 814.470034] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 814.616304] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464229, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.734330] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "a3e9af38-5115-4337-a20e-5d6c6ec655fe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.734330] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "a3e9af38-5115-4337-a20e-5d6c6ec655fe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.734566] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "a3e9af38-5115-4337-a20e-5d6c6ec655fe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.734666] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "a3e9af38-5115-4337-a20e-5d6c6ec655fe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.734846] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "a3e9af38-5115-4337-a20e-5d6c6ec655fe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.737401] env[61728]: INFO nova.compute.manager [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Terminating instance [ 814.739336] env[61728]: DEBUG nova.compute.manager [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 814.739545] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 814.740503] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-631de14c-a9c9-4a95-8bea-33fa1cc49465 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.748035] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 814.748329] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b18c6365-0e12-4db9-a6fd-4fe117505583 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.753913] env[61728]: DEBUG oslo_vmware.api [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 814.753913] env[61728]: value = "task-464232" [ 814.753913] env[61728]: _type = "Task" [ 814.753913] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.762049] env[61728]: DEBUG oslo_vmware.api [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-464232, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.789452] env[61728]: DEBUG oslo_vmware.api [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464230, 'name': PowerOffVM_Task, 'duration_secs': 0.307279} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.789990] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 814.790774] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 814.790774] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-90c8cae1-c304-4017-9c30-56825313803b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.845571] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464228, 'name': ReconfigVM_Task, 'duration_secs': 1.226149} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.846027] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Reconfigured VM instance instance-0000002b to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 814.846919] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fc5701e-a46f-41ea-a978-eee098e492ed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.849963] env[61728]: INFO nova.compute.manager [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Rebuilding instance [ 814.854362] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 814.854362] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 814.854362] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Deleting the datastore file [datastore1] b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 814.854362] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f775be4-b7a6-436d-9099-66beaf256876 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.879424] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] 1d9114cb-a0af-4874-962f-27237b3c89cc/1d9114cb-a0af-4874-962f-27237b3c89cc.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.883028] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2945da6-4d82-4197-88f2-f51d377c4b8f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.898476] env[61728]: DEBUG nova.scheduler.client.report [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 814.908124] env[61728]: DEBUG nova.network.neutron [None req-3590b696-2446-4ab2-9a45-b9d95d887cf5 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 814.910321] env[61728]: DEBUG oslo_vmware.api [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for the task: (returnval){ [ 814.910321] env[61728]: value = "task-464234" [ 814.910321] env[61728]: _type = "Task" [ 814.910321] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.921115] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 814.921115] env[61728]: value = "task-464235" [ 814.921115] env[61728]: _type = "Task" [ 814.921115] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.930479] env[61728]: DEBUG oslo_vmware.api [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464234, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.938440] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.939297] env[61728]: DEBUG nova.compute.manager [req-2dc3ec30-fb61-40da-a914-12acca2e351d req-c6150c21-18ec-4594-93d3-6ece20392811 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Received event network-changed-01099011-ff85-47eb-b05c-11566d3a6c0c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 814.943120] env[61728]: DEBUG nova.compute.manager [req-2dc3ec30-fb61-40da-a914-12acca2e351d req-c6150c21-18ec-4594-93d3-6ece20392811 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Refreshing instance network info cache due to event network-changed-01099011-ff85-47eb-b05c-11566d3a6c0c. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 814.943397] env[61728]: DEBUG oslo_concurrency.lockutils [req-2dc3ec30-fb61-40da-a914-12acca2e351d req-c6150c21-18ec-4594-93d3-6ece20392811 service nova] Acquiring lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.945943] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8136dc9-b7d2-4b08-addc-c7ff01ff7d4a tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Lock "6d55bb01-7855-4d86-ac57-ecae15ea590f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.457s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.956819] env[61728]: DEBUG nova.compute.manager [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.957782] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a78c7063-1652-4c96-9dfa-b33a95a62db2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.968155] env[61728]: DEBUG oslo_concurrency.lockutils [req-2bdc06fc-94b8-463c-a039-7e6840e7f06c req-0b64d157-e53f-40cb-a3af-ee125a76833f service nova] Releasing lock "refresh_cache-b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.974045] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 814.974045] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5296edfe-114b-8cda-26e6-2b2cb837079c" [ 814.974045] env[61728]: _type = "HttpNfcLease" [ 814.974045] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 814.974045] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 814.974045] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5296edfe-114b-8cda-26e6-2b2cb837079c" [ 814.974045] env[61728]: _type = "HttpNfcLease" [ 814.974045] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 814.974045] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318c3ff9-7a68-40a7-abb4-83c5e5167ee0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.981733] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52386a31-9571-1d1c-6c5a-ce204d25b9fa/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 814.981963] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52386a31-9571-1d1c-6c5a-ce204d25b9fa/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 815.081825] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fd3b44db-cd80-4e4b-9e38-343d6e8e583e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.121534] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464229, 'name': PowerOffVM_Task, 'duration_secs': 0.625867} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.121534] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 815.121873] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af3d727-f7d7-454e-911d-82fdbb7dbe29 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.141165] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b09eb5ed-0ef0-4426-ba3e-8cba6638225d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.263875] env[61728]: DEBUG oslo_vmware.api [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-464232, 'name': PowerOffVM_Task, 'duration_secs': 0.201026} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.264195] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 815.264458] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 815.264792] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a2416712-6235-4f6f-8678-82caad86c003 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.328304] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 815.328745] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 815.329419] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleting the datastore file [datastore1] a3e9af38-5115-4337-a20e-5d6c6ec655fe {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 815.329419] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-88c653da-daed-48d4-91dd-b2dbd325b997 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.337252] env[61728]: DEBUG oslo_vmware.api [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 815.337252] env[61728]: value = "task-464237" [ 815.337252] env[61728]: _type = "Task" [ 815.337252] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.345311] env[61728]: DEBUG oslo_vmware.api [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-464237, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.414478] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.150s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.416624] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 47.762s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.440129] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464235, 'name': ReconfigVM_Task, 'duration_secs': 0.460934} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.440835] env[61728]: DEBUG oslo_vmware.api [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Task: {'id': task-464234, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.268738} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.442037] env[61728]: INFO nova.scheduler.client.report [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Deleted allocations for instance 2e83145a-2481-411d-b037-a305a209074e [ 815.443287] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Reconfigured VM instance instance-0000002b to attach disk [datastore1] 1d9114cb-a0af-4874-962f-27237b3c89cc/1d9114cb-a0af-4874-962f-27237b3c89cc.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.444677] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 815.444917] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 815.445202] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 815.445411] env[61728]: INFO nova.compute.manager [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Took 1.19 seconds to destroy the instance on the hypervisor. [ 815.445852] env[61728]: DEBUG oslo.service.loopingcall [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.450820] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42bcdb7f-e99c-45a9-9317-031ca11de1d3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.453029] env[61728]: DEBUG nova.compute.manager [-] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.453269] env[61728]: DEBUG nova.network.neutron [-] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 815.480802] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 815.481301] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fbf26b4c-332a-49a0-b936-911a703979ab {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.484555] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2b03d5-3a1e-4189-a32d-95bde5d1b79e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.512544] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17812320-c41d-4d74-9084-e7a558c4c11f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.515660] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 815.515660] env[61728]: value = "task-464238" [ 815.515660] env[61728]: _type = "Task" [ 815.515660] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.539439] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b94fbf6d-7ace-49e0-8801-6a43be01cc10 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.546313] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464238, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.551955] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 815.552261] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-60f3cf69-62cb-4c56-befd-45998cf2c656 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.558342] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 815.558342] env[61728]: value = "task-464239" [ 815.558342] env[61728]: _type = "Task" [ 815.558342] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.569484] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464239, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.653298] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 815.653734] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b3514c58-9a39-45b7-849c-fd7abb8bc16f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.661289] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 815.661289] env[61728]: value = "task-464240" [ 815.661289] env[61728]: _type = "Task" [ 815.661289] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.672254] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464240, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.849244] env[61728]: DEBUG oslo_vmware.api [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-464237, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.21124} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.849675] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 815.849948] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 815.850326] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 815.850535] env[61728]: INFO nova.compute.manager [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Took 1.11 seconds to destroy the instance on the hypervisor. [ 815.851104] env[61728]: DEBUG oslo.service.loopingcall [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.851340] env[61728]: DEBUG nova.compute.manager [-] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.851922] env[61728]: DEBUG nova.network.neutron [-] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 815.958771] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7950df57-cbdd-47dc-b8b3-fa671f4fbb2d tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "2e83145a-2481-411d-b037-a305a209074e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 52.304s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.033105] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464238, 'name': PowerOffVM_Task, 'duration_secs': 0.176296} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.036666] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 816.036666] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 816.039728] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfca8190-0245-4f65-9eab-882f32cded57 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.058630] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 816.061772] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1bf98a3e-c2fd-4a67-a00c-7fab25d7d2be {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.073893] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464239, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.100236] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 816.100236] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 816.100236] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Deleting the datastore file [datastore1] 6d55bb01-7855-4d86-ac57-ecae15ea590f {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.100656] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fe8debc4-4a64-4b2f-aac2-710e03468395 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.111213] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 816.111213] env[61728]: value = "task-464242" [ 816.111213] env[61728]: _type = "Task" [ 816.111213] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.125912] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464242, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.175567] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464240, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.350008] env[61728]: DEBUG nova.network.neutron [None req-3590b696-2446-4ab2-9a45-b9d95d887cf5 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Updating instance_info_cache with network_info: [{"id": "01099011-ff85-47eb-b05c-11566d3a6c0c", "address": "fa:16:3e:a5:b6:18", "network": {"id": "02d4ac43-311d-4e89-ac44-d40b3790b873", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-928797810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f48655e35894d6daf7065aad4e84c2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01099011-ff", "ovs_interfaceid": "01099011-ff85-47eb-b05c-11566d3a6c0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.531205] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a7a444-10b8-4ebd-88dd-3515cc8d4637 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.539192] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb4b02d-e995-40f4-b23c-c3d53c9ca1c9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.573526] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8da2335-a0b0-4615-a538-70da0dfbab32 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.586268] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79843a8b-fb48-4df8-8b45-af8a3e8c53ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.590669] env[61728]: DEBUG oslo_vmware.api [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464239, 'name': PowerOnVM_Task, 'duration_secs': 0.565791} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.590954] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 816.608709] env[61728]: DEBUG nova.compute.provider_tree [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.621909] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464242, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.634360] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "dc8a587f-3dc8-43b3-b986-2096e3c4e1f1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.634616] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "dc8a587f-3dc8-43b3-b986-2096e3c4e1f1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.634852] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "dc8a587f-3dc8-43b3-b986-2096e3c4e1f1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.635051] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "dc8a587f-3dc8-43b3-b986-2096e3c4e1f1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.635235] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "dc8a587f-3dc8-43b3-b986-2096e3c4e1f1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.637354] env[61728]: INFO nova.compute.manager [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Terminating instance [ 816.639135] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "refresh_cache-dc8a587f-3dc8-43b3-b986-2096e3c4e1f1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.639301] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquired lock "refresh_cache-dc8a587f-3dc8-43b3-b986-2096e3c4e1f1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.639472] env[61728]: DEBUG nova.network.neutron [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 816.673603] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464240, 'name': CreateSnapshot_Task, 'duration_secs': 0.869434} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.673937] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 816.674840] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b59c5977-e1cc-4761-bcc3-2b8398c5e8bd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.854567] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3590b696-2446-4ab2-9a45-b9d95d887cf5 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Releasing lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.855466] env[61728]: DEBUG nova.compute.manager [None req-3590b696-2446-4ab2-9a45-b9d95d887cf5 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Inject network info {{(pid=61728) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 816.855466] env[61728]: DEBUG nova.compute.manager [None req-3590b696-2446-4ab2-9a45-b9d95d887cf5 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] network_info to inject: |[{"id": "01099011-ff85-47eb-b05c-11566d3a6c0c", "address": "fa:16:3e:a5:b6:18", "network": {"id": "02d4ac43-311d-4e89-ac44-d40b3790b873", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-928797810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f48655e35894d6daf7065aad4e84c2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01099011-ff", "ovs_interfaceid": "01099011-ff85-47eb-b05c-11566d3a6c0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 816.866577] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3590b696-2446-4ab2-9a45-b9d95d887cf5 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Reconfiguring VM instance to set the machine id {{(pid=61728) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 816.866577] env[61728]: DEBUG oslo_concurrency.lockutils [req-2dc3ec30-fb61-40da-a914-12acca2e351d req-c6150c21-18ec-4594-93d3-6ece20392811 service nova] Acquired lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.866577] env[61728]: DEBUG nova.network.neutron [req-2dc3ec30-fb61-40da-a914-12acca2e351d req-c6150c21-18ec-4594-93d3-6ece20392811 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Refreshing network info cache for port 01099011-ff85-47eb-b05c-11566d3a6c0c {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 816.868892] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df862eb4-8a13-4e14-900f-83ccac7073d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.886617] env[61728]: DEBUG oslo_vmware.api [None req-3590b696-2446-4ab2-9a45-b9d95d887cf5 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Waiting for the task: (returnval){ [ 816.886617] env[61728]: value = "task-464243" [ 816.886617] env[61728]: _type = "Task" [ 816.886617] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.899268] env[61728]: DEBUG nova.network.neutron [-] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.899268] env[61728]: DEBUG oslo_vmware.api [None req-3590b696-2446-4ab2-9a45-b9d95d887cf5 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-464243, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.968721] env[61728]: DEBUG nova.network.neutron [-] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.115060] env[61728]: DEBUG nova.scheduler.client.report [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 817.128765] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464242, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.696261} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.129071] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 817.129203] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 817.129410] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 817.138173] env[61728]: DEBUG nova.objects.instance [None req-31da54de-d023-40b9-a8c8-f2440f25986a tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Lazy-loading 'flavor' on Instance uuid e3116a87-ce12-4ec8-b6fa-70ec2127ebb0 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 817.159441] env[61728]: DEBUG nova.compute.manager [req-83185fd5-b58d-4f90-b508-6856a3e10018 req-408fa5ad-ebd4-4984-9581-a4f4b20bfc5e service nova] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Received event network-vif-deleted-920c9509-d0d8-464d-a5e2-bef1ba4f5637 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 817.159674] env[61728]: DEBUG nova.compute.manager [req-83185fd5-b58d-4f90-b508-6856a3e10018 req-408fa5ad-ebd4-4984-9581-a4f4b20bfc5e service nova] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Received event network-vif-deleted-0a12ba5f-fa23-4fa0-a00c-073b0f3cebc4 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 817.168631] env[61728]: DEBUG nova.network.neutron [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 817.195166] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 817.196063] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-abd53bff-4d13-4dd9-87ed-b19ee8d37f52 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.208077] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 817.208077] env[61728]: value = "task-464244" [ 817.208077] env[61728]: _type = "Task" [ 817.208077] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.217848] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464244, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.263720] env[61728]: DEBUG nova.network.neutron [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.405586] env[61728]: INFO nova.compute.manager [-] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Took 1.95 seconds to deallocate network for instance. [ 817.406947] env[61728]: DEBUG oslo_vmware.api [None req-3590b696-2446-4ab2-9a45-b9d95d887cf5 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-464243, 'name': ReconfigVM_Task, 'duration_secs': 0.177755} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.408729] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3590b696-2446-4ab2-9a45-b9d95d887cf5 tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Reconfigured VM instance to set the machine id {{(pid=61728) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 817.472211] env[61728]: INFO nova.compute.manager [-] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Took 1.62 seconds to deallocate network for instance. [ 817.605361] env[61728]: INFO nova.compute.manager [None req-5aa04767-521a-4ce8-ac15-4fdeeb40f7b5 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance to original state: 'active' [ 817.650618] env[61728]: DEBUG oslo_concurrency.lockutils [None req-31da54de-d023-40b9-a8c8-f2440f25986a tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquiring lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.720649] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464244, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.766635] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Releasing lock "refresh_cache-dc8a587f-3dc8-43b3-b986-2096e3c4e1f1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.767133] env[61728]: DEBUG nova.compute.manager [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 817.767357] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 817.768525] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c0afb04-39a8-4795-9eab-39d7419a8db2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.782519] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 817.783039] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ee689ab-09bb-4ced-a595-fc2ea01952db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.789908] env[61728]: DEBUG oslo_vmware.api [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 817.789908] env[61728]: value = "task-464245" [ 817.789908] env[61728]: _type = "Task" [ 817.789908] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.800975] env[61728]: DEBUG oslo_vmware.api [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464245, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.810806] env[61728]: DEBUG nova.network.neutron [req-2dc3ec30-fb61-40da-a914-12acca2e351d req-c6150c21-18ec-4594-93d3-6ece20392811 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Updated VIF entry in instance network info cache for port 01099011-ff85-47eb-b05c-11566d3a6c0c. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 817.811531] env[61728]: DEBUG nova.network.neutron [req-2dc3ec30-fb61-40da-a914-12acca2e351d req-c6150c21-18ec-4594-93d3-6ece20392811 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Updating instance_info_cache with network_info: [{"id": "01099011-ff85-47eb-b05c-11566d3a6c0c", "address": "fa:16:3e:a5:b6:18", "network": {"id": "02d4ac43-311d-4e89-ac44-d40b3790b873", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-928797810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f48655e35894d6daf7065aad4e84c2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01099011-ff", "ovs_interfaceid": "01099011-ff85-47eb-b05c-11566d3a6c0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.918853] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.979913] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.130513] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.714s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.130637] env[61728]: DEBUG nova.compute.manager [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61728) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 818.134719] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 49.279s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.176912] env[61728]: DEBUG nova.virt.hardware [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.177354] env[61728]: DEBUG nova.virt.hardware [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.177620] env[61728]: DEBUG nova.virt.hardware [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.177902] env[61728]: DEBUG nova.virt.hardware [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.178180] env[61728]: DEBUG nova.virt.hardware [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.178445] env[61728]: DEBUG nova.virt.hardware [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.178798] env[61728]: DEBUG nova.virt.hardware [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.179037] env[61728]: DEBUG nova.virt.hardware [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.179317] env[61728]: DEBUG nova.virt.hardware [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.179587] env[61728]: DEBUG nova.virt.hardware [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.179894] env[61728]: DEBUG nova.virt.hardware [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.181121] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e93d08f4-cb14-4bd8-8a04-c74f08fd5823 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.195798] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513ba188-610f-4f07-a427-3705947ecfc2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.214472] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Instance VIF info [] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.220857] env[61728]: DEBUG oslo.service.loopingcall [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.224937] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 818.225291] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3285af2e-d751-4d45-b34b-9d2586dba07a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.244817] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464244, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.246453] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.246453] env[61728]: value = "task-464246" [ 818.246453] env[61728]: _type = "Task" [ 818.246453] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.260972] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464246, 'name': CreateVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.278276] env[61728]: DEBUG oslo_concurrency.lockutils [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Acquiring lock "a5c7e3ff-6829-4528-9279-e865a0eb8512" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.278613] env[61728]: DEBUG oslo_concurrency.lockutils [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Lock "a5c7e3ff-6829-4528-9279-e865a0eb8512" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.278915] env[61728]: DEBUG oslo_concurrency.lockutils [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Acquiring lock "a5c7e3ff-6829-4528-9279-e865a0eb8512-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.279428] env[61728]: DEBUG oslo_concurrency.lockutils [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Lock "a5c7e3ff-6829-4528-9279-e865a0eb8512-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.279500] env[61728]: DEBUG oslo_concurrency.lockutils [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Lock "a5c7e3ff-6829-4528-9279-e865a0eb8512-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.282903] env[61728]: INFO nova.compute.manager [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Terminating instance [ 818.287020] env[61728]: DEBUG nova.compute.manager [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 818.287020] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 818.287367] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3f2a801-9912-4102-aca3-369d6102d1d9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.299194] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 818.303657] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e3c90d7-7fbd-47bb-bb7d-b7ef7b7ad1bb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.305790] env[61728]: DEBUG oslo_vmware.api [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464245, 'name': PowerOffVM_Task, 'duration_secs': 0.155968} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.306211] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 818.306475] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 818.307352] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1aa6d5a-7445-4eb6-aea3-0c424f08845b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.314463] env[61728]: DEBUG oslo_concurrency.lockutils [req-2dc3ec30-fb61-40da-a914-12acca2e351d req-c6150c21-18ec-4594-93d3-6ece20392811 service nova] Releasing lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.315819] env[61728]: DEBUG oslo_vmware.api [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Waiting for the task: (returnval){ [ 818.315819] env[61728]: value = "task-464247" [ 818.315819] env[61728]: _type = "Task" [ 818.315819] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.315819] env[61728]: DEBUG oslo_concurrency.lockutils [None req-31da54de-d023-40b9-a8c8-f2440f25986a tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquired lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.328210] env[61728]: DEBUG oslo_vmware.api [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464247, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.339657] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 818.339657] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 818.339860] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Deleting the datastore file [datastore1] dc8a587f-3dc8-43b3-b986-2096e3c4e1f1 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 818.340181] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-309019c0-f98d-46ef-b053-aa000de3a971 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.348284] env[61728]: DEBUG oslo_vmware.api [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for the task: (returnval){ [ 818.348284] env[61728]: value = "task-464249" [ 818.348284] env[61728]: _type = "Task" [ 818.348284] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.358349] env[61728]: DEBUG oslo_vmware.api [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464249, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.555027] env[61728]: DEBUG oslo_concurrency.lockutils [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.555027] env[61728]: DEBUG oslo_concurrency.lockutils [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.724844] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464244, 'name': CloneVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.729145] env[61728]: INFO nova.scheduler.client.report [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleted allocation for migration 7ca3d98f-8659-4e9a-bbb2-d0f80e820da2 [ 818.756318] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464246, 'name': CreateVM_Task, 'duration_secs': 0.401769} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.756521] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 818.757226] env[61728]: DEBUG oslo_concurrency.lockutils [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.757374] env[61728]: DEBUG oslo_concurrency.lockutils [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.757743] env[61728]: DEBUG oslo_concurrency.lockutils [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 818.758112] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebc1854a-6af1-46bc-a567-614624e679f7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.764331] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 818.764331] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52406512-ec3f-41c5-487d-c17dfdac6134" [ 818.764331] env[61728]: _type = "Task" [ 818.764331] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.775569] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52406512-ec3f-41c5-487d-c17dfdac6134, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.827926] env[61728]: DEBUG oslo_vmware.api [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464247, 'name': PowerOffVM_Task, 'duration_secs': 0.178462} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.829115] env[61728]: DEBUG nova.network.neutron [None req-31da54de-d023-40b9-a8c8-f2440f25986a tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 818.830686] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 818.831463] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 818.832076] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b8c12443-2dd6-4ca6-b63c-8388b117a222 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.858422] env[61728]: DEBUG oslo_vmware.api [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Task: {'id': task-464249, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169459} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.858761] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 818.859063] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 818.859341] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 818.859562] env[61728]: INFO nova.compute.manager [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Took 1.09 seconds to destroy the instance on the hypervisor. [ 818.859942] env[61728]: DEBUG oslo.service.loopingcall [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.860417] env[61728]: DEBUG nova.compute.manager [-] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 818.860709] env[61728]: DEBUG nova.network.neutron [-] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 818.884390] env[61728]: DEBUG nova.network.neutron [-] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 818.896695] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 818.896960] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 818.897210] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Deleting the datastore file [datastore1] a5c7e3ff-6829-4528-9279-e865a0eb8512 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 818.897788] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8aa1a655-1672-42ab-a40e-9ade19f00f31 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.904491] env[61728]: DEBUG oslo_vmware.api [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Waiting for the task: (returnval){ [ 818.904491] env[61728]: value = "task-464251" [ 818.904491] env[61728]: _type = "Task" [ 818.904491] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.914035] env[61728]: DEBUG oslo_vmware.api [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464251, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.057983] env[61728]: DEBUG nova.compute.utils [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.158044] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Applying migration context for instance f2692e22-6c14-4bb3-a080-607f0731105d as it has an incoming, in-progress migration adc30b71-8803-41e0-a72a-09cb8bee068c. Migration status is confirming {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 819.160198] env[61728]: INFO nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating resource usage from migration adc30b71-8803-41e0-a72a-09cb8bee068c [ 819.187290] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 771341ed-8b8e-470c-9686-82650f5271b4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.187480] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance a3e9af38-5115-4337-a20e-5d6c6ec655fe is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 819.187662] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.187730] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 4e5dc620-68a1-4e10-a8be-702c4999ca10 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.187855] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 37b8ca5a-e0f8-414b-a363-e56db520f027 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 819.187985] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 819.188115] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance e3116a87-ce12-4ec8-b6fa-70ec2127ebb0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.188241] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 0fb1192e-99f1-4469-b196-60df7eab8185 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 819.188359] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 3bfefcfc-db97-4a9d-86cb-9fb1d8158863 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.188474] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.188645] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 6df7b619-8cc4-4dd9-8596-22dc83234a8e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 819.188707] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance c84bfcd0-b145-4675-8b0a-5e8f94f65098 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.188820] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance dc8a587f-3dc8-43b3-b986-2096e3c4e1f1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.188930] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 40ef29df-12d6-4274-846a-4a2c32752b8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.189048] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 1d9114cb-a0af-4874-962f-27237b3c89cc actively managed on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.189181] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 39eaa6f4-df3f-4928-a1f1-8c861a392e29 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 819.189314] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance b63fa3df-317a-46d4-b8e9-74e9e287efde is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 819.189436] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 06e765e1-85d9-4f6c-8ab5-81735f08cb66 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.189550] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance a5c7e3ff-6829-4528-9279-e865a0eb8512 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.189754] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.189885] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance f28348d2-c062-497a-b374-521df51054ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.190024] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance a716cf5a-4482-45cb-96ec-b8c38bc3e742 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 819.190148] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 819.190264] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 06109957-2c3b-404e-a43e-dd34ece39096 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.190376] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 2976d71f-7322-4640-b734-645dbb70e0c3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.190488] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Migration adc30b71-8803-41e0-a72a-09cb8bee068c is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 819.190598] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance f2692e22-6c14-4bb3-a080-607f0731105d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.190707] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 6d55bb01-7855-4d86-ac57-ecae15ea590f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.224432] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464244, 'name': CloneVM_Task, 'duration_secs': 1.582777} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.224733] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Created linked-clone VM from snapshot [ 819.225525] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8aa908-345e-4c93-a108-642b180e0f17 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.234779] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Uploading image fbcbbf71-c4ec-41db-9a9c-0ca05fc90c2f {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 819.237809] env[61728]: DEBUG oslo_concurrency.lockutils [None req-69b84014-501f-4f8f-8276-48d9e2825def tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 54.668s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.265034] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 819.265034] env[61728]: value = "vm-122113" [ 819.265034] env[61728]: _type = "VirtualMachine" [ 819.265034] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 819.265386] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ad53cf3d-1e1c-4641-bdb6-74026636f3d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.276912] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52406512-ec3f-41c5-487d-c17dfdac6134, 'name': SearchDatastore_Task, 'duration_secs': 0.02559} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.277726] env[61728]: DEBUG oslo_concurrency.lockutils [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.278084] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 819.278414] env[61728]: DEBUG oslo_concurrency.lockutils [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.278585] env[61728]: DEBUG oslo_concurrency.lockutils [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.278778] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 819.279160] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lease: (returnval){ [ 819.279160] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b20a92-d809-e1e0-3368-9a2b70c86859" [ 819.279160] env[61728]: _type = "HttpNfcLease" [ 819.279160] env[61728]: } obtained for exporting VM: (result){ [ 819.279160] env[61728]: value = "vm-122113" [ 819.279160] env[61728]: _type = "VirtualMachine" [ 819.279160] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 819.279395] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the lease: (returnval){ [ 819.279395] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b20a92-d809-e1e0-3368-9a2b70c86859" [ 819.279395] env[61728]: _type = "HttpNfcLease" [ 819.279395] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 819.279588] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e63d0d94-673d-4e33-ab59-24a084ddcc42 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.289094] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 819.289094] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b20a92-d809-e1e0-3368-9a2b70c86859" [ 819.289094] env[61728]: _type = "HttpNfcLease" [ 819.289094] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 819.291826] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 819.292034] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 819.292776] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d018109-361a-41d6-ba25-1657986d0710 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.297644] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 819.297644] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523abda0-b723-9509-84e3-549cc591685c" [ 819.297644] env[61728]: _type = "Task" [ 819.297644] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.305184] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523abda0-b723-9509-84e3-549cc591685c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.388103] env[61728]: DEBUG nova.network.neutron [-] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.419735] env[61728]: DEBUG oslo_vmware.api [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Task: {'id': task-464251, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228602} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.420011] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 819.420224] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 819.420414] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 819.420598] env[61728]: INFO nova.compute.manager [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Took 1.14 seconds to destroy the instance on the hypervisor. [ 819.420856] env[61728]: DEBUG oslo.service.loopingcall [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.421072] env[61728]: DEBUG nova.compute.manager [-] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 819.421200] env[61728]: DEBUG nova.network.neutron [-] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 819.491537] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "1d9114cb-a0af-4874-962f-27237b3c89cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.491880] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "1d9114cb-a0af-4874-962f-27237b3c89cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.492122] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "1d9114cb-a0af-4874-962f-27237b3c89cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.492318] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "1d9114cb-a0af-4874-962f-27237b3c89cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.492491] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "1d9114cb-a0af-4874-962f-27237b3c89cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.494748] env[61728]: INFO nova.compute.manager [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Terminating instance [ 819.496791] env[61728]: DEBUG nova.compute.manager [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 819.497006] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 819.497836] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a4931c-8dc9-469d-bbd8-43a9bf969d6e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.506504] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 819.506674] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e246bb9-5d65-4ef3-8f0b-30c2d0bb6d7f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.513846] env[61728]: DEBUG oslo_vmware.api [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 819.513846] env[61728]: value = "task-464253" [ 819.513846] env[61728]: _type = "Task" [ 819.513846] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.531307] env[61728]: DEBUG oslo_vmware.api [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464253, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.561351] env[61728]: DEBUG oslo_concurrency.lockutils [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.693976] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 3ef9bf60-c6a1-4b7a-a375-5397fb871850 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 819.719931] env[61728]: DEBUG nova.network.neutron [None req-31da54de-d023-40b9-a8c8-f2440f25986a tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Updating instance_info_cache with network_info: [{"id": "01099011-ff85-47eb-b05c-11566d3a6c0c", "address": "fa:16:3e:a5:b6:18", "network": {"id": "02d4ac43-311d-4e89-ac44-d40b3790b873", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-928797810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f48655e35894d6daf7065aad4e84c2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01099011-ff", "ovs_interfaceid": "01099011-ff85-47eb-b05c-11566d3a6c0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.790185] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 819.790185] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b20a92-d809-e1e0-3368-9a2b70c86859" [ 819.790185] env[61728]: _type = "HttpNfcLease" [ 819.790185] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 819.792018] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 819.792018] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b20a92-d809-e1e0-3368-9a2b70c86859" [ 819.792018] env[61728]: _type = "HttpNfcLease" [ 819.792018] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 819.792018] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2632a6-5a26-4cb2-9678-c066b8091322 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.803296] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ce2fdf-5b53-5253-65b7-ca24142f8a25/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 819.803296] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ce2fdf-5b53-5253-65b7-ca24142f8a25/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 819.875776] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523abda0-b723-9509-84e3-549cc591685c, 'name': SearchDatastore_Task, 'duration_secs': 0.011782} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.880797] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe0f1da5-ca6b-4438-92fa-d55438bde218 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.888130] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 819.888130] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524f6cd6-fd36-cef5-d321-ae13c41c661a" [ 819.888130] env[61728]: _type = "Task" [ 819.888130] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.893479] env[61728]: INFO nova.compute.manager [-] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Took 1.03 seconds to deallocate network for instance. [ 819.910443] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524f6cd6-fd36-cef5-d321-ae13c41c661a, 'name': SearchDatastore_Task, 'duration_secs': 0.013314} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.913116] env[61728]: DEBUG oslo_concurrency.lockutils [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.913434] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 6d55bb01-7855-4d86-ac57-ecae15ea590f/6d55bb01-7855-4d86-ac57-ecae15ea590f.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 819.914152] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9603049-29d6-48db-a2ae-8fb207ea0d4f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.916433] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c6fcd98e-ca28-4580-87f8-ae093083f183 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.925708] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 819.925708] env[61728]: value = "task-464254" [ 819.925708] env[61728]: _type = "Task" [ 819.925708] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.931363] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464254, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.954279] env[61728]: DEBUG nova.objects.instance [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'flavor' on Instance uuid 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 820.026346] env[61728]: DEBUG oslo_vmware.api [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464253, 'name': PowerOffVM_Task, 'duration_secs': 0.18851} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.026614] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 820.026874] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 820.027371] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f369ef6a-bccb-4eb1-93b6-a1fd328d2f61 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.197372] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance d44718bd-872b-401c-aa11-f10bea4a35d8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 820.222999] env[61728]: DEBUG oslo_concurrency.lockutils [None req-31da54de-d023-40b9-a8c8-f2440f25986a tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Releasing lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.222999] env[61728]: DEBUG nova.compute.manager [None req-31da54de-d023-40b9-a8c8-f2440f25986a tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Inject network info {{(pid=61728) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 820.223252] env[61728]: DEBUG nova.compute.manager [None req-31da54de-d023-40b9-a8c8-f2440f25986a tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] network_info to inject: |[{"id": "01099011-ff85-47eb-b05c-11566d3a6c0c", "address": "fa:16:3e:a5:b6:18", "network": {"id": "02d4ac43-311d-4e89-ac44-d40b3790b873", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-928797810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f48655e35894d6daf7065aad4e84c2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01099011-ff", "ovs_interfaceid": "01099011-ff85-47eb-b05c-11566d3a6c0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 820.231028] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-31da54de-d023-40b9-a8c8-f2440f25986a tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Reconfiguring VM instance to set the machine id {{(pid=61728) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 820.231028] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1aa9a2b0-2e04-4d9d-8ab6-32af7d618ec2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.249575] env[61728]: DEBUG oslo_vmware.api [None req-31da54de-d023-40b9-a8c8-f2440f25986a tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Waiting for the task: (returnval){ [ 820.249575] env[61728]: value = "task-464256" [ 820.249575] env[61728]: _type = "Task" [ 820.249575] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.264735] env[61728]: DEBUG oslo_vmware.api [None req-31da54de-d023-40b9-a8c8-f2440f25986a tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-464256, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.269847] env[61728]: DEBUG nova.compute.manager [req-3bb3221a-81c2-4192-901c-e82517e775ca req-4c2c377c-92ba-4f36-87e3-16d9e9aeeaf5 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Received event network-changed-01099011-ff85-47eb-b05c-11566d3a6c0c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 820.269847] env[61728]: DEBUG nova.compute.manager [req-3bb3221a-81c2-4192-901c-e82517e775ca req-4c2c377c-92ba-4f36-87e3-16d9e9aeeaf5 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Refreshing instance network info cache due to event network-changed-01099011-ff85-47eb-b05c-11566d3a6c0c. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 820.269847] env[61728]: DEBUG oslo_concurrency.lockutils [req-3bb3221a-81c2-4192-901c-e82517e775ca req-4c2c377c-92ba-4f36-87e3-16d9e9aeeaf5 service nova] Acquiring lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.269847] env[61728]: DEBUG oslo_concurrency.lockutils [req-3bb3221a-81c2-4192-901c-e82517e775ca req-4c2c377c-92ba-4f36-87e3-16d9e9aeeaf5 service nova] Acquired lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.269847] env[61728]: DEBUG nova.network.neutron [req-3bb3221a-81c2-4192-901c-e82517e775ca req-4c2c377c-92ba-4f36-87e3-16d9e9aeeaf5 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Refreshing network info cache for port 01099011-ff85-47eb-b05c-11566d3a6c0c {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 820.329874] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 820.329874] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 820.329874] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Deleting the datastore file [datastore1] 1d9114cb-a0af-4874-962f-27237b3c89cc {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 820.330384] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6cf0ca55-082d-4410-8b24-c1c9a09cdeeb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.341946] env[61728]: DEBUG oslo_vmware.api [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 820.341946] env[61728]: value = "task-464257" [ 820.341946] env[61728]: _type = "Task" [ 820.341946] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.352503] env[61728]: DEBUG oslo_vmware.api [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464257, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.403727] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.435237] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464254, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.463216] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.463416] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.463602] env[61728]: DEBUG nova.network.neutron [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 820.463811] env[61728]: DEBUG nova.objects.instance [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'info_cache' on Instance uuid 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 820.522658] env[61728]: DEBUG nova.network.neutron [-] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.666123] env[61728]: DEBUG oslo_concurrency.lockutils [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.666123] env[61728]: DEBUG oslo_concurrency.lockutils [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.666425] env[61728]: INFO nova.compute.manager [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Attaching volume 0835974c-5660-4c88-a603-e2bd917f000b to /dev/sdb [ 820.700658] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance b444a6cf-3c98-4308-afc1-6e760d30082d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 820.713433] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bdfd73e-9612-4437-8709-d3f7dc287fe1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.723756] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6585727-605f-463b-9193-86916c0f9ba4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.740537] env[61728]: DEBUG nova.virt.block_device [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Updating existing volume attachment record: 1c536f50-5063-4b91-a060-a7ac83524ed4 {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 820.759051] env[61728]: DEBUG oslo_vmware.api [None req-31da54de-d023-40b9-a8c8-f2440f25986a tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-464256, 'name': ReconfigVM_Task, 'duration_secs': 0.206826} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.759529] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-31da54de-d023-40b9-a8c8-f2440f25986a tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Reconfigured VM instance to set the machine id {{(pid=61728) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 820.854673] env[61728]: DEBUG oslo_vmware.api [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464257, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.392221} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.854673] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 820.855501] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 820.856162] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 820.856664] env[61728]: INFO nova.compute.manager [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Took 1.36 seconds to destroy the instance on the hypervisor. [ 820.856931] env[61728]: DEBUG oslo.service.loopingcall [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.857213] env[61728]: DEBUG nova.compute.manager [-] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 820.858785] env[61728]: DEBUG nova.network.neutron [-] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 820.938990] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464254, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.644293} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.940151] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 6d55bb01-7855-4d86-ac57-ecae15ea590f/6d55bb01-7855-4d86-ac57-ecae15ea590f.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 820.942187] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.942187] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bef83b4e-2b3b-4725-abf9-efedfde417f3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.951206] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 820.951206] env[61728]: value = "task-464259" [ 820.951206] env[61728]: _type = "Task" [ 820.951206] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.962539] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464259, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.969694] env[61728]: DEBUG nova.objects.base [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Object Instance<7b16fbc0-7f13-405f-b84e-e18de1ca7dd2> lazy-loaded attributes: flavor,info_cache {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 820.979426] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Acquiring lock "a9745dc8-6400-424c-bc10-1181ee128648" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.980573] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Lock "a9745dc8-6400-424c-bc10-1181ee128648" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.027010] env[61728]: INFO nova.compute.manager [-] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Took 1.60 seconds to deallocate network for instance. [ 821.198077] env[61728]: DEBUG nova.network.neutron [req-3bb3221a-81c2-4192-901c-e82517e775ca req-4c2c377c-92ba-4f36-87e3-16d9e9aeeaf5 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Updated VIF entry in instance network info cache for port 01099011-ff85-47eb-b05c-11566d3a6c0c. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 821.198561] env[61728]: DEBUG nova.network.neutron [req-3bb3221a-81c2-4192-901c-e82517e775ca req-4c2c377c-92ba-4f36-87e3-16d9e9aeeaf5 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Updating instance_info_cache with network_info: [{"id": "01099011-ff85-47eb-b05c-11566d3a6c0c", "address": "fa:16:3e:a5:b6:18", "network": {"id": "02d4ac43-311d-4e89-ac44-d40b3790b873", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-928797810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f48655e35894d6daf7065aad4e84c2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01099011-ff", "ovs_interfaceid": "01099011-ff85-47eb-b05c-11566d3a6c0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.205999] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 174607a1-9bc8-4e07-8993-7f0bb0f308e9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 821.206451] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Total usable vcpus: 48, total allocated vcpus: 19 {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 821.206611] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4352MB phys_disk=200GB used_disk=19GB total_vcpus=48 used_vcpus=19 pci_stats=[] {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 821.469718] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464259, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073948} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.473027] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 821.474984] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3abf40cc-bb8f-4dcd-a3d9-d003e15af7eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.493519] env[61728]: DEBUG nova.compute.manager [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 821.507181] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 6d55bb01-7855-4d86-ac57-ecae15ea590f/6d55bb01-7855-4d86-ac57-ecae15ea590f.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.512602] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d117ea9-221f-4012-a995-4f4dbd481c2d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.537047] env[61728]: DEBUG oslo_concurrency.lockutils [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.540264] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 821.540264] env[61728]: value = "task-464262" [ 821.540264] env[61728]: _type = "Task" [ 821.540264] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.548349] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464262, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.595156] env[61728]: DEBUG oslo_concurrency.lockutils [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquiring lock "e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.595646] env[61728]: DEBUG oslo_concurrency.lockutils [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Lock "e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.595908] env[61728]: DEBUG oslo_concurrency.lockutils [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquiring lock "e3116a87-ce12-4ec8-b6fa-70ec2127ebb0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.596156] env[61728]: DEBUG oslo_concurrency.lockutils [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Lock "e3116a87-ce12-4ec8-b6fa-70ec2127ebb0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.596348] env[61728]: DEBUG oslo_concurrency.lockutils [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Lock "e3116a87-ce12-4ec8-b6fa-70ec2127ebb0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.599789] env[61728]: INFO nova.compute.manager [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Terminating instance [ 821.602550] env[61728]: DEBUG nova.compute.manager [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 821.602907] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 821.603921] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283831ab-42ec-4a61-b791-867e25192c3e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.616945] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 821.617300] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ac0bcad-cf2b-4bdb-9333-95d02cedcf7d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.627762] env[61728]: DEBUG oslo_vmware.api [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Waiting for the task: (returnval){ [ 821.627762] env[61728]: value = "task-464263" [ 821.627762] env[61728]: _type = "Task" [ 821.627762] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.637261] env[61728]: DEBUG oslo_vmware.api [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-464263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.706559] env[61728]: DEBUG oslo_concurrency.lockutils [req-3bb3221a-81c2-4192-901c-e82517e775ca req-4c2c377c-92ba-4f36-87e3-16d9e9aeeaf5 service nova] Releasing lock "refresh_cache-e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.799954] env[61728]: DEBUG nova.network.neutron [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance_info_cache with network_info: [{"id": "6c9a6b61-5248-4996-a390-c18edb3ec021", "address": "fa:16:3e:43:9a:d5", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9a6b61-52", "ovs_interfaceid": "6c9a6b61-5248-4996-a390-c18edb3ec021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.828567] env[61728]: DEBUG nova.network.neutron [-] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.857446] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9acac501-22cc-4560-9c68-437eeb4f7c95 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.866032] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7dbfd5-f0ef-4578-933a-6073272c9f31 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.897381] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd191aa8-d197-43a4-ae4a-b9cb28026662 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.905388] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba6ea97-3bfe-4ca3-ba1e-295df966db14 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.919170] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.030990] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.048853] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464262, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.136534] env[61728]: DEBUG oslo_vmware.api [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-464263, 'name': PowerOffVM_Task, 'duration_secs': 0.200166} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.136834] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 822.137017] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 822.137292] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-578a128e-46a9-4f2e-b2ea-f9a2b80d6c91 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.302546] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "refresh_cache-7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.316871] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 822.317331] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 822.317495] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Deleting the datastore file [datastore1] e3116a87-ce12-4ec8-b6fa-70ec2127ebb0 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 822.317847] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95ac36b6-0b3e-49ea-9f58-50f62474c8d2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.325639] env[61728]: DEBUG oslo_vmware.api [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Waiting for the task: (returnval){ [ 822.325639] env[61728]: value = "task-464265" [ 822.325639] env[61728]: _type = "Task" [ 822.325639] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.334318] env[61728]: INFO nova.compute.manager [-] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Took 1.48 seconds to deallocate network for instance. [ 822.335038] env[61728]: DEBUG oslo_vmware.api [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-464265, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.422611] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.549295] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464262, 'name': ReconfigVM_Task, 'duration_secs': 0.530897} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.549698] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 6d55bb01-7855-4d86-ac57-ecae15ea590f/6d55bb01-7855-4d86-ac57-ecae15ea590f.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 822.550416] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e773a0f2-6991-49bb-b619-91707a2a4c56 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.556671] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 822.556671] env[61728]: value = "task-464266" [ 822.556671] env[61728]: _type = "Task" [ 822.556671] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.564956] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464266, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.809825] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 822.810274] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e531eb6e-7dbf-4112-9689-4cd237835602 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.818833] env[61728]: DEBUG oslo_vmware.api [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 822.818833] env[61728]: value = "task-464267" [ 822.818833] env[61728]: _type = "Task" [ 822.818833] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.832806] env[61728]: DEBUG oslo_vmware.api [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464267, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.839017] env[61728]: DEBUG oslo_vmware.api [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Task: {'id': task-464265, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203524} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.839350] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 822.839568] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 822.840916] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 822.840916] env[61728]: INFO nova.compute.manager [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Took 1.24 seconds to destroy the instance on the hypervisor. [ 822.840916] env[61728]: DEBUG oslo.service.loopingcall [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.840916] env[61728]: DEBUG nova.compute.manager [-] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 822.840916] env[61728]: DEBUG nova.network.neutron [-] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 822.843144] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.928777] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 822.931686] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.795s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.931863] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 51.268s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.933496] env[61728]: INFO nova.compute.claims [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.939451] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 822.939451] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Cleaning up deleted instances {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 822.991049] env[61728]: DEBUG nova.compute.manager [req-0d80d775-9135-4805-9525-d7fc8d7e41d2 req-977df1af-1f9a-4a3f-8782-7ee3647ce5b5 service nova] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Received event network-vif-deleted-6f900c64-beea-4e8a-84c5-9f13768f8420 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 822.991049] env[61728]: DEBUG nova.compute.manager [req-0d80d775-9135-4805-9525-d7fc8d7e41d2 req-977df1af-1f9a-4a3f-8782-7ee3647ce5b5 service nova] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Received event network-vif-deleted-4d70757b-d260-4aa3-8914-a38eb567f7b8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 823.072590] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464266, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.327682] env[61728]: DEBUG oslo_vmware.api [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464267, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.464887] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] There are 39 instances to clean {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 823.466059] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 2e83145a-2481-411d-b037-a305a209074e] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 823.568293] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464266, 'name': Rename_Task, 'duration_secs': 0.660481} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.568636] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 823.568902] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b48df6a1-3631-4bc7-a6eb-405d7319b816 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.576215] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 823.576215] env[61728]: value = "task-464269" [ 823.576215] env[61728]: _type = "Task" [ 823.576215] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.585367] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464269, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.766385] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52386a31-9571-1d1c-6c5a-ce204d25b9fa/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 823.766904] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde5a235-9287-4a87-829c-010d13fc0789 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.774194] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52386a31-9571-1d1c-6c5a-ce204d25b9fa/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 823.774409] env[61728]: ERROR oslo_vmware.rw_handles [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52386a31-9571-1d1c-6c5a-ce204d25b9fa/disk-0.vmdk due to incomplete transfer. [ 823.774927] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-11b7c5f7-c439-40fd-a30a-b3fd4b94c206 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.782178] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52386a31-9571-1d1c-6c5a-ce204d25b9fa/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 823.782397] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Uploaded image 82a9852c-0eb3-4602-b891-4c6c26b90c26 to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 823.784727] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 823.785028] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d41b2422-d036-4f66-a394-b45d66eba678 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.790732] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 823.790732] env[61728]: value = "task-464270" [ 823.790732] env[61728]: _type = "Task" [ 823.790732] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.801877] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464270, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.829064] env[61728]: DEBUG oslo_vmware.api [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464267, 'name': PowerOnVM_Task, 'duration_secs': 0.672791} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.829367] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 823.829635] env[61728]: DEBUG nova.compute.manager [None req-e7522cc9-2554-4b54-9c1b-30738884dd79 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.830487] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412d8fc8-e30f-48fa-9353-2e1307bee361 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.975512] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 3fc8bcc5-3dd0-491a-8a3e-4b9d317d8e48] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 824.053154] env[61728]: DEBUG nova.network.neutron [-] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.094525] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464269, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.304213] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464270, 'name': Destroy_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.483149] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 223b1f93-d33e-4566-8c8e-75234a491814] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 824.560152] env[61728]: INFO nova.compute.manager [-] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Took 1.72 seconds to deallocate network for instance. [ 824.591710] env[61728]: DEBUG oslo_vmware.api [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464269, 'name': PowerOnVM_Task, 'duration_secs': 0.783966} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.594961] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 824.594961] env[61728]: DEBUG nova.compute.manager [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 824.595605] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dae77fd-11da-44f3-b431-5da6cea16d86 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.636180] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e09da0-988d-44b9-9e36-2be562bb57ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.648972] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16d2066-46e9-48a4-81b0-39ef889737e6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.695874] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8290e8-f40a-43a0-8808-f8b8e7076679 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.700739] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48caef3-c2e0-42f3-a9ba-2a1a726c53a4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.714954] env[61728]: DEBUG nova.compute.provider_tree [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 824.802596] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464270, 'name': Destroy_Task, 'duration_secs': 0.561867} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.802979] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Destroyed the VM [ 824.803174] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 824.803500] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6bb361a0-9d00-4988-a09a-2768607260f1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.811178] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 824.811178] env[61728]: value = "task-464271" [ 824.811178] env[61728]: _type = "Task" [ 824.811178] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.819576] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464271, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.983443] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: c14034b5-330b-4fb1-af31-d530e3231992] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 825.070709] env[61728]: DEBUG oslo_concurrency.lockutils [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.115165] env[61728]: DEBUG oslo_concurrency.lockutils [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.245279] env[61728]: ERROR nova.scheduler.client.report [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [req-feb1e616-3314-4947-b113-860f70a72b7a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-feb1e616-3314-4947-b113-860f70a72b7a"}]} [ 825.267894] env[61728]: DEBUG nova.scheduler.client.report [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 825.293855] env[61728]: DEBUG nova.scheduler.client.report [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 825.294551] env[61728]: DEBUG nova.compute.provider_tree [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 825.302917] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "d8657743-30f9-4f8f-aca8-d98f3c557d19" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.302917] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "d8657743-30f9-4f8f-aca8-d98f3c557d19" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.317157] env[61728]: DEBUG nova.scheduler.client.report [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 825.326129] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464271, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.343041] env[61728]: DEBUG nova.scheduler.client.report [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 825.487587] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: b563e521-0e0e-45ea-bd5a-056285ea35b9] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 825.566199] env[61728]: DEBUG nova.compute.manager [req-b0d9cd39-18b5-4fc4-824f-f3904353ad68 req-c3c2b19f-5317-4ec7-964e-2eafe650e6b3 service nova] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Received event network-vif-deleted-01099011-ff85-47eb-b05c-11566d3a6c0c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 825.580691] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.580691] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.580691] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.580909] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.580981] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.584254] env[61728]: INFO nova.compute.manager [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Terminating instance [ 825.586225] env[61728]: DEBUG nova.compute.manager [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 825.586427] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 825.587298] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48465cb1-d8c2-46a3-81ac-2f9e37647a94 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.601064] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 825.601064] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-07c34bf8-9148-44f4-9ba2-7760053bb8fe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.607348] env[61728]: DEBUG oslo_vmware.api [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 825.607348] env[61728]: value = "task-464272" [ 825.607348] env[61728]: _type = "Task" [ 825.607348] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.616628] env[61728]: DEBUG oslo_vmware.api [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464272, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.806159] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 825.806973] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122116', 'volume_id': '0835974c-5660-4c88-a603-e2bd917f000b', 'name': 'volume-0835974c-5660-4c88-a603-e2bd917f000b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '06e765e1-85d9-4f6c-8ab5-81735f08cb66', 'attached_at': '', 'detached_at': '', 'volume_id': '0835974c-5660-4c88-a603-e2bd917f000b', 'serial': '0835974c-5660-4c88-a603-e2bd917f000b'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 825.807779] env[61728]: DEBUG nova.compute.manager [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.811036] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06994786-3292-4b95-a656-854396f01ab1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.837078] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af4fbff-f25b-402d-8612-c4ec957c9e1a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.843048] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464271, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.866589] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] volume-0835974c-5660-4c88-a603-e2bd917f000b/volume-0835974c-5660-4c88-a603-e2bd917f000b.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.868417] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88bed39f-f621-4d49-88d3-e7717c846a94 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.883402] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f846997-63a9-4a5e-b4ca-728c141191df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.894040] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ea88377-152a-4fe3-b245-0bf088c861ab {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.897635] env[61728]: DEBUG oslo_vmware.api [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 825.897635] env[61728]: value = "task-464273" [ 825.897635] env[61728]: _type = "Task" [ 825.897635] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.930934] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b04db28-bec9-4208-a4c3-747c3114bbbc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.937298] env[61728]: DEBUG oslo_vmware.api [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.942756] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e267ed9b-2950-4946-8760-d0304ddbf2cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.962280] env[61728]: DEBUG nova.compute.provider_tree [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 825.991883] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: ca8d41dc-5719-4ade-b82a-b1d7f1c8def1] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 826.117285] env[61728]: DEBUG oslo_vmware.api [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464272, 'name': PowerOffVM_Task, 'duration_secs': 0.300613} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.117647] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 826.117884] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 826.118182] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f831263c-8a80-4e60-9406-6a045ec2790c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.185848] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 826.186182] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 826.186452] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleting the datastore file [datastore1] 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 826.186774] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f13a0ed-bdee-4810-8287-af86e2fb9c52 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.194110] env[61728]: DEBUG oslo_vmware.api [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 826.194110] env[61728]: value = "task-464275" [ 826.194110] env[61728]: _type = "Task" [ 826.194110] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.202587] env[61728]: DEBUG oslo_vmware.api [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464275, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.329664] env[61728]: DEBUG oslo_vmware.api [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464271, 'name': RemoveSnapshot_Task, 'duration_secs': 1.232772} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.329664] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 826.329873] env[61728]: INFO nova.compute.manager [None req-d18209c2-def1-42f6-b227-346522995519 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Took 17.05 seconds to snapshot the instance on the hypervisor. [ 826.339978] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.412483] env[61728]: DEBUG oslo_vmware.api [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.495201] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: b833f691-e76e-4a2e-94a4-7594fadc3ba8] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 826.502984] env[61728]: DEBUG nova.scheduler.client.report [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 107 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 826.503506] env[61728]: DEBUG nova.compute.provider_tree [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 107 to 108 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 826.503506] env[61728]: DEBUG nova.compute.provider_tree [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 826.642040] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6bc2efdd-62e9-4653-99d9-6f2dd1f03a67 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "interface-f28348d2-c062-497a-b374-521df51054ee-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.642260] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6bc2efdd-62e9-4653-99d9-6f2dd1f03a67 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-f28348d2-c062-497a-b374-521df51054ee-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.642634] env[61728]: DEBUG nova.objects.instance [None req-6bc2efdd-62e9-4653-99d9-6f2dd1f03a67 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'flavor' on Instance uuid f28348d2-c062-497a-b374-521df51054ee {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 826.704377] env[61728]: DEBUG oslo_vmware.api [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464275, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.405192} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.704652] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 826.704848] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 826.705079] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 826.705269] env[61728]: INFO nova.compute.manager [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Took 1.12 seconds to destroy the instance on the hypervisor. [ 826.705516] env[61728]: DEBUG oslo.service.loopingcall [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.705713] env[61728]: DEBUG nova.compute.manager [-] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 826.705829] env[61728]: DEBUG nova.network.neutron [-] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 826.908213] env[61728]: DEBUG oslo_concurrency.lockutils [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquiring lock "6d55bb01-7855-4d86-ac57-ecae15ea590f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.908557] env[61728]: DEBUG oslo_concurrency.lockutils [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Lock "6d55bb01-7855-4d86-ac57-ecae15ea590f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.908864] env[61728]: DEBUG oslo_concurrency.lockutils [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquiring lock "6d55bb01-7855-4d86-ac57-ecae15ea590f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.909114] env[61728]: DEBUG oslo_concurrency.lockutils [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Lock "6d55bb01-7855-4d86-ac57-ecae15ea590f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.909291] env[61728]: DEBUG oslo_concurrency.lockutils [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Lock "6d55bb01-7855-4d86-ac57-ecae15ea590f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.911206] env[61728]: DEBUG oslo_vmware.api [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464273, 'name': ReconfigVM_Task, 'duration_secs': 0.761812} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.911764] env[61728]: INFO nova.compute.manager [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Terminating instance [ 826.913370] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Reconfigured VM instance instance-0000003b to attach disk [datastore1] volume-0835974c-5660-4c88-a603-e2bd917f000b/volume-0835974c-5660-4c88-a603-e2bd917f000b.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.918663] env[61728]: DEBUG oslo_concurrency.lockutils [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquiring lock "refresh_cache-6d55bb01-7855-4d86-ac57-ecae15ea590f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.918880] env[61728]: DEBUG oslo_concurrency.lockutils [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquired lock "refresh_cache-6d55bb01-7855-4d86-ac57-ecae15ea590f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.919102] env[61728]: DEBUG nova.network.neutron [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 826.920387] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b48365b-cc9f-4aa8-8095-ab4a699fbc14 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.939702] env[61728]: DEBUG oslo_vmware.api [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 826.939702] env[61728]: value = "task-464276" [ 826.939702] env[61728]: _type = "Task" [ 826.939702] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.951019] env[61728]: DEBUG oslo_vmware.api [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464276, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.953711] env[61728]: DEBUG nova.network.neutron [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 827.002784] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: ad304c11-4ff8-4910-92a5-24216d727c4b] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 827.008775] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.077s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.009411] env[61728]: DEBUG nova.compute.manager [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 827.012651] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 53.879s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.014226] env[61728]: INFO nova.compute.claims [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.044468] env[61728]: DEBUG nova.network.neutron [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.148090] env[61728]: DEBUG nova.objects.instance [None req-6bc2efdd-62e9-4653-99d9-6f2dd1f03a67 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'pci_requests' on Instance uuid f28348d2-c062-497a-b374-521df51054ee {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.452315] env[61728]: DEBUG oslo_vmware.api [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464276, 'name': ReconfigVM_Task, 'duration_secs': 0.152961} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.452315] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122116', 'volume_id': '0835974c-5660-4c88-a603-e2bd917f000b', 'name': 'volume-0835974c-5660-4c88-a603-e2bd917f000b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '06e765e1-85d9-4f6c-8ab5-81735f08cb66', 'attached_at': '', 'detached_at': '', 'volume_id': '0835974c-5660-4c88-a603-e2bd917f000b', 'serial': '0835974c-5660-4c88-a603-e2bd917f000b'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 827.506803] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 650d2875-68f7-4131-b9b7-9799ee08f1d6] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 827.521528] env[61728]: DEBUG nova.compute.utils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.523462] env[61728]: DEBUG nova.compute.manager [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 827.523592] env[61728]: DEBUG nova.network.neutron [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 827.547363] env[61728]: DEBUG oslo_concurrency.lockutils [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Releasing lock "refresh_cache-6d55bb01-7855-4d86-ac57-ecae15ea590f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.547830] env[61728]: DEBUG nova.compute.manager [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 827.548039] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 827.549168] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb916c4-d388-4c9d-b119-b6bbfd91dc3e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.557351] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 827.558094] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93ad3c1b-4446-43af-a1f7-e08464867433 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.563706] env[61728]: DEBUG oslo_vmware.api [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 827.563706] env[61728]: value = "task-464277" [ 827.563706] env[61728]: _type = "Task" [ 827.563706] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.565245] env[61728]: DEBUG nova.policy [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aac08b5321344d5ab384a3714e6e0bc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '383f38592de14b4ba93fa9a6eeec40aa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 827.575442] env[61728]: DEBUG oslo_vmware.api [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464277, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.650904] env[61728]: DEBUG nova.objects.base [None req-6bc2efdd-62e9-4653-99d9-6f2dd1f03a67 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 827.651148] env[61728]: DEBUG nova.network.neutron [None req-6bc2efdd-62e9-4653-99d9-6f2dd1f03a67 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 827.737841] env[61728]: DEBUG nova.network.neutron [-] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.789082] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6bc2efdd-62e9-4653-99d9-6f2dd1f03a67 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-f28348d2-c062-497a-b374-521df51054ee-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.147s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.873895] env[61728]: DEBUG nova.compute.manager [req-0d0ab814-849a-4693-b783-5f9587113eab req-8134014f-9f77-4f8a-88ce-30bb5ec5fb8f service nova] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Received event network-vif-deleted-6c9a6b61-5248-4996-a390-c18edb3ec021 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 827.895801] env[61728]: DEBUG nova.network.neutron [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Successfully created port: ebbb109a-f3c6-4a01-a162-7d8bca2c868c {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 828.010358] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 507b5aba-aec9-438b-92c0-5efa17959bc7] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 828.028153] env[61728]: DEBUG nova.compute.manager [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 828.077472] env[61728]: DEBUG oslo_vmware.api [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464277, 'name': PowerOffVM_Task, 'duration_secs': 0.127741} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.080174] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 828.080364] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 828.081380] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49047347-8672-4bda-8ec1-2f66fb5b283d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.107055] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 828.107055] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 828.107055] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Deleting the datastore file [datastore1] 6d55bb01-7855-4d86-ac57-ecae15ea590f {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 828.107055] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-60253a3c-29ec-48a0-898c-05f082c4bea9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.114178] env[61728]: DEBUG oslo_vmware.api [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for the task: (returnval){ [ 828.114178] env[61728]: value = "task-464279" [ 828.114178] env[61728]: _type = "Task" [ 828.114178] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.124612] env[61728]: DEBUG oslo_vmware.api [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464279, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.237522] env[61728]: INFO nova.compute.manager [-] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Took 1.53 seconds to deallocate network for instance. [ 828.489738] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba1bec7-2599-4e08-b4f3-a6c92cded788 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.497647] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3dab3ed-4864-4047-ab1a-ea0c2bbef037 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.533598] env[61728]: DEBUG nova.objects.instance [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lazy-loading 'flavor' on Instance uuid 06e765e1-85d9-4f6c-8ab5-81735f08cb66 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 828.535084] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 3395fc8a-eefa-4e87-9eb0-f61b0c0da0a1] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 828.541313] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a27b12f-e2b3-4c9a-af85-9a165f9324c0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.557813] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5978ea61-3956-44e7-abde-9a73921d96d1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.574818] env[61728]: DEBUG nova.compute.provider_tree [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.577722] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ce2fdf-5b53-5253-65b7-ca24142f8a25/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 828.578775] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-444de51f-5122-471a-97aa-8cb9093b7381 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.584872] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ce2fdf-5b53-5253-65b7-ca24142f8a25/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 828.585056] env[61728]: ERROR oslo_vmware.rw_handles [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ce2fdf-5b53-5253-65b7-ca24142f8a25/disk-0.vmdk due to incomplete transfer. [ 828.585276] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f45fe60d-2944-419f-b706-8af5822d40a3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.591821] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ce2fdf-5b53-5253-65b7-ca24142f8a25/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 828.592050] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Uploaded image fbcbbf71-c4ec-41db-9a9c-0ca05fc90c2f to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 828.593693] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 828.593962] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d8b4ad17-d39c-493d-a664-af4d1515762d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.599346] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 828.599346] env[61728]: value = "task-464280" [ 828.599346] env[61728]: _type = "Task" [ 828.599346] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.606640] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464280, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.620963] env[61728]: DEBUG oslo_vmware.api [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Task: {'id': task-464279, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.112552} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.621258] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 828.621514] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 828.621668] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 828.621888] env[61728]: INFO nova.compute.manager [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Took 1.07 seconds to destroy the instance on the hypervisor. [ 828.622185] env[61728]: DEBUG oslo.service.loopingcall [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.622407] env[61728]: DEBUG nova.compute.manager [-] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 828.622503] env[61728]: DEBUG nova.network.neutron [-] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 828.637905] env[61728]: DEBUG nova.network.neutron [-] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 828.747042] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.038639] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 8c53c9c9-cdc1-407b-8827-8409cf137235] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 829.043026] env[61728]: DEBUG nova.compute.manager [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 829.045343] env[61728]: DEBUG oslo_concurrency.lockutils [None req-28ae4f61-bdf2-4eaa-9736-1ed056bfaf45 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.379s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.074120] env[61728]: DEBUG nova.virt.hardware [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 829.074397] env[61728]: DEBUG nova.virt.hardware [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 829.074597] env[61728]: DEBUG nova.virt.hardware [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.074806] env[61728]: DEBUG nova.virt.hardware [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 829.074965] env[61728]: DEBUG nova.virt.hardware [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.075146] env[61728]: DEBUG nova.virt.hardware [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 829.075361] env[61728]: DEBUG nova.virt.hardware [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 829.075653] env[61728]: DEBUG nova.virt.hardware [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 829.075849] env[61728]: DEBUG nova.virt.hardware [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 829.076075] env[61728]: DEBUG nova.virt.hardware [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 829.076291] env[61728]: DEBUG nova.virt.hardware [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.077399] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd27b4d-0065-4e9b-8efc-d3907a9868ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.083138] env[61728]: DEBUG nova.scheduler.client.report [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.090147] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae17400-b8b9-4bb3-a58d-867beca2f29b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.112907] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464280, 'name': Destroy_Task, 'duration_secs': 0.318967} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.113197] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Destroyed the VM [ 829.113437] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 829.113684] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a67e7e20-df9d-4efd-baf5-8572bc0433fa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.120786] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 829.120786] env[61728]: value = "task-464281" [ 829.120786] env[61728]: _type = "Task" [ 829.120786] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.129279] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464281, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.140945] env[61728]: DEBUG nova.network.neutron [-] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.542841] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 69d7af16-7708-4df6-acca-093b6c1df1ba] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 829.589847] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.577s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.590397] env[61728]: DEBUG nova.compute.manager [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 829.597486] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 54.507s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.597688] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.602228] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 53.565s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.602228] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.605419] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 52.074s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.605614] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.607258] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 49.411s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.607451] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.609029] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 45.377s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.610659] env[61728]: INFO nova.compute.claims [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.620016] env[61728]: DEBUG nova.network.neutron [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Successfully updated port: ebbb109a-f3c6-4a01-a162-7d8bca2c868c {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.633650] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464281, 'name': RemoveSnapshot_Task, 'duration_secs': 0.38327} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.635071] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 829.636966] env[61728]: DEBUG nova.compute.manager [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.636966] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c268b0d-479d-49c5-a020-c9ab289e6405 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.641121] env[61728]: INFO nova.scheduler.client.report [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Deleted allocations for instance b63fa3df-317a-46d4-b8e9-74e9e287efde [ 829.641822] env[61728]: INFO nova.scheduler.client.report [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Deleted allocations for instance 37b8ca5a-e0f8-414b-a363-e56db520f027 [ 829.647384] env[61728]: INFO nova.compute.manager [-] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Took 1.02 seconds to deallocate network for instance. [ 829.661859] env[61728]: INFO nova.scheduler.client.report [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Deleted allocations for instance 39eaa6f4-df3f-4928-a1f1-8c861a392e29 [ 829.667742] env[61728]: INFO nova.scheduler.client.report [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Deleted allocations for instance c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df [ 829.671645] env[61728]: WARNING oslo_messaging._drivers.amqpdriver [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 829.927613] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "interface-f28348d2-c062-497a-b374-521df51054ee-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.927893] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-f28348d2-c062-497a-b374-521df51054ee-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.928260] env[61728]: DEBUG nova.objects.instance [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'flavor' on Instance uuid f28348d2-c062-497a-b374-521df51054ee {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 830.045878] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 4fcc957e-5175-4589-b4da-d8451d92c815] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 830.116668] env[61728]: DEBUG nova.compute.utils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 830.120729] env[61728]: DEBUG nova.compute.manager [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 830.120961] env[61728]: DEBUG nova.network.neutron [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 830.123892] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "refresh_cache-3ef9bf60-c6a1-4b7a-a375-5397fb871850" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.123972] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquired lock "refresh_cache-3ef9bf60-c6a1-4b7a-a375-5397fb871850" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.124111] env[61728]: DEBUG nova.network.neutron [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 830.164100] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e04dd4dc-5d53-40fe-a60f-d24f2497b76f tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "b63fa3df-317a-46d4-b8e9-74e9e287efde" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 56.022s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.165219] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ae46a04c-54ad-4bf8-aa19-543df6c3efa2 tempest-ServersTestBootFromVolume-909910790 tempest-ServersTestBootFromVolume-909910790-project-member] Lock "37b8ca5a-e0f8-414b-a363-e56db520f027" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 59.388s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.171583] env[61728]: DEBUG nova.policy [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54c30aeeea3a4c1b8ed34430361c2a4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a02bd852c564acd99e12fd17279f028', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 830.177528] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a84436d-a043-44a1-bea7-ae3dd5b9c099 tempest-SecurityGroupsTestJSON-1032747602 tempest-SecurityGroupsTestJSON-1032747602-project-member] Lock "c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 53.172s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.178492] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f7775580-e407-40d1-8f72-c131b32f2d31 tempest-ServerGroupTestJSON-398825368 tempest-ServerGroupTestJSON-398825368-project-member] Lock "39eaa6f4-df3f-4928-a1f1-8c861a392e29" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 58.924s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.179804] env[61728]: DEBUG oslo_concurrency.lockutils [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.182236] env[61728]: INFO nova.compute.manager [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Shelve offloading [ 830.183951] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 830.184222] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9af61fd2-bc22-4535-8251-7a60568f40df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.192519] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 830.192519] env[61728]: value = "task-464282" [ 830.192519] env[61728]: _type = "Task" [ 830.192519] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.200631] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464282, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.418325] env[61728]: DEBUG nova.network.neutron [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Successfully created port: 1a5d5fc4-4ae6-467a-8bff-0c0f410c9252 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 830.549322] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 7a60fcd0-01b0-40bc-93e7-6f5ce03232d4] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 830.572862] env[61728]: DEBUG nova.objects.instance [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'pci_requests' on Instance uuid f28348d2-c062-497a-b374-521df51054ee {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 830.577556] env[61728]: DEBUG nova.compute.manager [req-2063608e-6b32-4893-967e-fccc65a6a8bc req-e3844b60-50e8-47e8-9ec1-ad0fef6a72bb service nova] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Received event network-vif-plugged-ebbb109a-f3c6-4a01-a162-7d8bca2c868c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.577786] env[61728]: DEBUG oslo_concurrency.lockutils [req-2063608e-6b32-4893-967e-fccc65a6a8bc req-e3844b60-50e8-47e8-9ec1-ad0fef6a72bb service nova] Acquiring lock "3ef9bf60-c6a1-4b7a-a375-5397fb871850-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.578047] env[61728]: DEBUG oslo_concurrency.lockutils [req-2063608e-6b32-4893-967e-fccc65a6a8bc req-e3844b60-50e8-47e8-9ec1-ad0fef6a72bb service nova] Lock "3ef9bf60-c6a1-4b7a-a375-5397fb871850-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.578213] env[61728]: DEBUG oslo_concurrency.lockutils [req-2063608e-6b32-4893-967e-fccc65a6a8bc req-e3844b60-50e8-47e8-9ec1-ad0fef6a72bb service nova] Lock "3ef9bf60-c6a1-4b7a-a375-5397fb871850-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.578351] env[61728]: DEBUG nova.compute.manager [req-2063608e-6b32-4893-967e-fccc65a6a8bc req-e3844b60-50e8-47e8-9ec1-ad0fef6a72bb service nova] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] No waiting events found dispatching network-vif-plugged-ebbb109a-f3c6-4a01-a162-7d8bca2c868c {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 830.578523] env[61728]: WARNING nova.compute.manager [req-2063608e-6b32-4893-967e-fccc65a6a8bc req-e3844b60-50e8-47e8-9ec1-ad0fef6a72bb service nova] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Received unexpected event network-vif-plugged-ebbb109a-f3c6-4a01-a162-7d8bca2c868c for instance with vm_state building and task_state spawning. [ 830.578688] env[61728]: DEBUG nova.compute.manager [req-2063608e-6b32-4893-967e-fccc65a6a8bc req-e3844b60-50e8-47e8-9ec1-ad0fef6a72bb service nova] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Received event network-changed-ebbb109a-f3c6-4a01-a162-7d8bca2c868c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.578877] env[61728]: DEBUG nova.compute.manager [req-2063608e-6b32-4893-967e-fccc65a6a8bc req-e3844b60-50e8-47e8-9ec1-ad0fef6a72bb service nova] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Refreshing instance network info cache due to event network-changed-ebbb109a-f3c6-4a01-a162-7d8bca2c868c. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 830.579571] env[61728]: DEBUG oslo_concurrency.lockutils [req-2063608e-6b32-4893-967e-fccc65a6a8bc req-e3844b60-50e8-47e8-9ec1-ad0fef6a72bb service nova] Acquiring lock "refresh_cache-3ef9bf60-c6a1-4b7a-a375-5397fb871850" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.627466] env[61728]: DEBUG nova.compute.manager [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 830.678304] env[61728]: DEBUG nova.network.neutron [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 830.711396] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 830.711679] env[61728]: DEBUG nova.compute.manager [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.712846] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a49739-6b40-4ef4-b757-c78df718ec9e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.724061] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "refresh_cache-2976d71f-7322-4640-b734-645dbb70e0c3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.727027] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "refresh_cache-2976d71f-7322-4640-b734-645dbb70e0c3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.727027] env[61728]: DEBUG nova.network.neutron [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 830.868706] env[61728]: DEBUG nova.network.neutron [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Updating instance_info_cache with network_info: [{"id": "ebbb109a-f3c6-4a01-a162-7d8bca2c868c", "address": "fa:16:3e:78:7b:b0", "network": {"id": "17a8ea01-7320-4089-9138-a6fd17ac4054", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-599506352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "383f38592de14b4ba93fa9a6eeec40aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebbb109a-f3", "ovs_interfaceid": "ebbb109a-f3c6-4a01-a162-7d8bca2c868c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.053823] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 1ef96606-d960-4283-b372-e2bad47f9367] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 831.076782] env[61728]: DEBUG nova.objects.base [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 831.077241] env[61728]: DEBUG nova.network.neutron [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 831.079876] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4edd55d5-6dec-4c86-a6bc-9064efc2a85f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.089788] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85a844c-3ca8-48e5-96b5-8aae7adf5b13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.144850] env[61728]: DEBUG nova.policy [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e51efcf23db4d5496edbc3b7104b4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32bb35e1dfce40e48be08bb568d3f2b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 831.148036] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f2bcdb-8dff-4924-bf68-a847e758a38a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.160270] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24bb8155-c3ea-45b2-885d-62aa1709457c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.185737] env[61728]: DEBUG nova.compute.provider_tree [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.193429] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "87540771-c864-4d49-8ad3-b6b559841761" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.193802] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "87540771-c864-4d49-8ad3-b6b559841761" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.371867] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Releasing lock "refresh_cache-3ef9bf60-c6a1-4b7a-a375-5397fb871850" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.372252] env[61728]: DEBUG nova.compute.manager [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Instance network_info: |[{"id": "ebbb109a-f3c6-4a01-a162-7d8bca2c868c", "address": "fa:16:3e:78:7b:b0", "network": {"id": "17a8ea01-7320-4089-9138-a6fd17ac4054", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-599506352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "383f38592de14b4ba93fa9a6eeec40aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebbb109a-f3", "ovs_interfaceid": "ebbb109a-f3c6-4a01-a162-7d8bca2c868c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 831.372593] env[61728]: DEBUG oslo_concurrency.lockutils [req-2063608e-6b32-4893-967e-fccc65a6a8bc req-e3844b60-50e8-47e8-9ec1-ad0fef6a72bb service nova] Acquired lock "refresh_cache-3ef9bf60-c6a1-4b7a-a375-5397fb871850" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.372817] env[61728]: DEBUG nova.network.neutron [req-2063608e-6b32-4893-967e-fccc65a6a8bc req-e3844b60-50e8-47e8-9ec1-ad0fef6a72bb service nova] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Refreshing network info cache for port ebbb109a-f3c6-4a01-a162-7d8bca2c868c {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 831.374112] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:7b:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ebbb109a-f3c6-4a01-a162-7d8bca2c868c', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 831.381774] env[61728]: DEBUG oslo.service.loopingcall [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.384740] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 831.386069] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aff5ff35-2a8e-4780-bd24-b92f671f6d29 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.411822] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 831.411822] env[61728]: value = "task-464283" [ 831.411822] env[61728]: _type = "Task" [ 831.411822] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.416249] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464283, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.557701] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: fbd521c8-9618-4b5c-839e-50d8631fa7b9] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 831.654946] env[61728]: DEBUG nova.compute.manager [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 831.691772] env[61728]: DEBUG nova.virt.hardware [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.692137] env[61728]: DEBUG nova.virt.hardware [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.692326] env[61728]: DEBUG nova.virt.hardware [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.692525] env[61728]: DEBUG nova.virt.hardware [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.692722] env[61728]: DEBUG nova.virt.hardware [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.692907] env[61728]: DEBUG nova.virt.hardware [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.693175] env[61728]: DEBUG nova.virt.hardware [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.693367] env[61728]: DEBUG nova.virt.hardware [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.693553] env[61728]: DEBUG nova.virt.hardware [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.693739] env[61728]: DEBUG nova.virt.hardware [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.694371] env[61728]: DEBUG nova.virt.hardware [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.695347] env[61728]: DEBUG nova.scheduler.client.report [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.702654] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396958e5-52d7-4a34-8c82-89a39b98d384 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.703012] env[61728]: DEBUG nova.compute.manager [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 831.712903] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2965e80-cd57-4e09-b5d8-bb2b20cc2c5f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.734505] env[61728]: DEBUG nova.network.neutron [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Updating instance_info_cache with network_info: [{"id": "78aa30fc-a52b-4ebd-bb26-1020af74cf2e", "address": "fa:16:3e:16:c0:48", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78aa30fc-a5", "ovs_interfaceid": "78aa30fc-a52b-4ebd-bb26-1020af74cf2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.791900] env[61728]: DEBUG nova.network.neutron [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Successfully created port: 02293cc2-3b2d-4b93-b27d-0e61b2780e89 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 831.921427] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464283, 'name': CreateVM_Task, 'duration_secs': 0.303718} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.921721] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 831.922403] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.922573] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.922955] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.923247] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64f3f06f-a782-45a4-99cc-6933acd2f687 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.928744] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 831.928744] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52634660-5632-99bf-a8d5-ff783ebedbd9" [ 831.928744] env[61728]: _type = "Task" [ 831.928744] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.936754] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52634660-5632-99bf-a8d5-ff783ebedbd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.064612] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 0f75629b-bdec-4593-aceb-cd0478972ccc] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 832.203658] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.594s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.204237] env[61728]: DEBUG nova.compute.manager [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 832.207213] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.818s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.207413] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.209564] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.416s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.211949] env[61728]: INFO nova.compute.claims [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 832.233999] env[61728]: INFO nova.scheduler.client.report [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Deleted allocations for instance 6df7b619-8cc4-4dd9-8596-22dc83234a8e [ 832.239266] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "refresh_cache-2976d71f-7322-4640-b734-645dbb70e0c3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.245797] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.441081] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52634660-5632-99bf-a8d5-ff783ebedbd9, 'name': SearchDatastore_Task, 'duration_secs': 0.016391} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.441683] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.442046] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.442316] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.442470] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.442676] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.442970] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2df8a9f1-aecd-4f4f-8657-fea9dbf17eb9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.455358] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.455548] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 832.456499] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f04949c3-fdf6-4c65-bef7-b24d23be88e7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.462593] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 832.462593] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d3514f-7e97-af4e-13a7-80a66551c263" [ 832.462593] env[61728]: _type = "Task" [ 832.462593] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.471082] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d3514f-7e97-af4e-13a7-80a66551c263, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.494064] env[61728]: DEBUG nova.network.neutron [req-2063608e-6b32-4893-967e-fccc65a6a8bc req-e3844b60-50e8-47e8-9ec1-ad0fef6a72bb service nova] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Updated VIF entry in instance network info cache for port ebbb109a-f3c6-4a01-a162-7d8bca2c868c. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 832.494483] env[61728]: DEBUG nova.network.neutron [req-2063608e-6b32-4893-967e-fccc65a6a8bc req-e3844b60-50e8-47e8-9ec1-ad0fef6a72bb service nova] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Updating instance_info_cache with network_info: [{"id": "ebbb109a-f3c6-4a01-a162-7d8bca2c868c", "address": "fa:16:3e:78:7b:b0", "network": {"id": "17a8ea01-7320-4089-9138-a6fd17ac4054", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-599506352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "383f38592de14b4ba93fa9a6eeec40aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebbb109a-f3", "ovs_interfaceid": "ebbb109a-f3c6-4a01-a162-7d8bca2c868c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.556274] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 832.557194] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f171f613-a3c5-4862-b3d5-f5303a0af6f6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.565437] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 832.565714] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22283a37-dee5-4e7f-88de-ba1735006805 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.567449] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f947d808-ebea-4593-b8e6-d3851c5216dc] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 832.634223] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 832.634512] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 832.634640] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleting the datastore file [datastore1] 2976d71f-7322-4640-b734-645dbb70e0c3 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.634939] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be41fd1f-ca03-49ba-854d-4667fc9cba6b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.642494] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 832.642494] env[61728]: value = "task-464285" [ 832.642494] env[61728]: _type = "Task" [ 832.642494] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.650329] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464285, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.716584] env[61728]: DEBUG nova.compute.utils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 832.718546] env[61728]: DEBUG nova.compute.manager [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 832.718748] env[61728]: DEBUG nova.network.neutron [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 832.745440] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1c1d0ea4-1b00-401f-951c-8f5ea98aa394 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "6df7b619-8cc4-4dd9-8596-22dc83234a8e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 43.333s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.775841] env[61728]: DEBUG nova.policy [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd53ec83e72e44563946aa13db7e7ca02', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '29dfac4a253f43be811df957e0ee5f96', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 832.818848] env[61728]: DEBUG nova.network.neutron [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Successfully updated port: 1a5d5fc4-4ae6-467a-8bff-0c0f410c9252 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 832.977207] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d3514f-7e97-af4e-13a7-80a66551c263, 'name': SearchDatastore_Task, 'duration_secs': 0.018618} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.978334] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62fb58ca-99f0-4fa0-8bbe-321495324c48 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.983777] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 832.983777] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5286cd6f-f961-bd1d-3250-8db13dc961f7" [ 832.983777] env[61728]: _type = "Task" [ 832.983777] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.992986] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5286cd6f-f961-bd1d-3250-8db13dc961f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.997429] env[61728]: DEBUG oslo_concurrency.lockutils [req-2063608e-6b32-4893-967e-fccc65a6a8bc req-e3844b60-50e8-47e8-9ec1-ad0fef6a72bb service nova] Releasing lock "refresh_cache-3ef9bf60-c6a1-4b7a-a375-5397fb871850" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.070367] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: fc3c1d93-b095-4d80-8d71-243b66d85a46] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 833.152555] env[61728]: DEBUG oslo_vmware.api [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464285, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155932} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.153476] env[61728]: DEBUG nova.network.neutron [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Successfully created port: 8904a7fb-a87a-4f80-958e-87748f00b59e {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 833.155727] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 833.155935] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 833.156128] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 833.162571] env[61728]: DEBUG oslo_concurrency.lockutils [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Acquiring lock "4e5dc620-68a1-4e10-a8be-702c4999ca10" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.163075] env[61728]: DEBUG oslo_concurrency.lockutils [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Lock "4e5dc620-68a1-4e10-a8be-702c4999ca10" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.163169] env[61728]: DEBUG oslo_concurrency.lockutils [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Acquiring lock "4e5dc620-68a1-4e10-a8be-702c4999ca10-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.164142] env[61728]: DEBUG oslo_concurrency.lockutils [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Lock "4e5dc620-68a1-4e10-a8be-702c4999ca10-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.164142] env[61728]: DEBUG oslo_concurrency.lockutils [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Lock "4e5dc620-68a1-4e10-a8be-702c4999ca10-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.165975] env[61728]: INFO nova.compute.manager [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Terminating instance [ 833.167856] env[61728]: DEBUG nova.compute.manager [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 833.168063] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 833.169216] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1319234-e9c9-44ae-83d3-dffd5e7a0379 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.179696] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 833.180659] env[61728]: INFO nova.scheduler.client.report [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleted allocations for instance 2976d71f-7322-4640-b734-645dbb70e0c3 [ 833.183242] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3649e777-a684-4084-a8b1-301ef2c6c6d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.190482] env[61728]: DEBUG oslo_vmware.api [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Waiting for the task: (returnval){ [ 833.190482] env[61728]: value = "task-464286" [ 833.190482] env[61728]: _type = "Task" [ 833.190482] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.198469] env[61728]: DEBUG oslo_vmware.api [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-464286, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.224182] env[61728]: DEBUG nova.compute.manager [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 833.326028] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "refresh_cache-d44718bd-872b-401c-aa11-f10bea4a35d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.326028] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "refresh_cache-d44718bd-872b-401c-aa11-f10bea4a35d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.326028] env[61728]: DEBUG nova.network.neutron [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 833.497123] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5286cd6f-f961-bd1d-3250-8db13dc961f7, 'name': SearchDatastore_Task, 'duration_secs': 0.008992} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.497466] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.498123] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3ef9bf60-c6a1-4b7a-a375-5397fb871850/3ef9bf60-c6a1-4b7a-a375-5397fb871850.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 833.498381] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ead9d709-2b48-4020-8f93-1d62707a565c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.509299] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 833.509299] env[61728]: value = "task-464287" [ 833.509299] env[61728]: _type = "Task" [ 833.509299] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.518027] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464287, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.574886] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 2d389d5e-a53f-424e-90af-5ed05e2f93c7] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 833.686393] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.697934] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8969fefa-6d54-4477-98ca-d5316a9e0aff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.706871] env[61728]: DEBUG oslo_vmware.api [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-464286, 'name': PowerOffVM_Task, 'duration_secs': 0.191602} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.707641] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 833.708224] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 833.708542] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-82992d08-60eb-48af-8f59-e3ec79a58cde {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.713627] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a4b4ed-fcbe-44fa-9527-79a5087f1773 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.751815] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08132534-eeb2-4c45-9c6d-2f23a8b25755 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.761686] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6086166a-5acd-489a-b7da-297aabec9060 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.779882] env[61728]: DEBUG nova.compute.provider_tree [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.791495] env[61728]: DEBUG nova.compute.manager [req-257c616e-80f8-4b74-b522-8a19313b0f31 req-bf6f176e-94f7-42f8-8a07-85a3f38a9724 service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Received event network-vif-unplugged-78aa30fc-a52b-4ebd-bb26-1020af74cf2e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 833.791584] env[61728]: DEBUG oslo_concurrency.lockutils [req-257c616e-80f8-4b74-b522-8a19313b0f31 req-bf6f176e-94f7-42f8-8a07-85a3f38a9724 service nova] Acquiring lock "2976d71f-7322-4640-b734-645dbb70e0c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.792597] env[61728]: DEBUG oslo_concurrency.lockutils [req-257c616e-80f8-4b74-b522-8a19313b0f31 req-bf6f176e-94f7-42f8-8a07-85a3f38a9724 service nova] Lock "2976d71f-7322-4640-b734-645dbb70e0c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.792597] env[61728]: DEBUG oslo_concurrency.lockutils [req-257c616e-80f8-4b74-b522-8a19313b0f31 req-bf6f176e-94f7-42f8-8a07-85a3f38a9724 service nova] Lock "2976d71f-7322-4640-b734-645dbb70e0c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.792597] env[61728]: DEBUG nova.compute.manager [req-257c616e-80f8-4b74-b522-8a19313b0f31 req-bf6f176e-94f7-42f8-8a07-85a3f38a9724 service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] No waiting events found dispatching network-vif-unplugged-78aa30fc-a52b-4ebd-bb26-1020af74cf2e {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 833.792597] env[61728]: WARNING nova.compute.manager [req-257c616e-80f8-4b74-b522-8a19313b0f31 req-bf6f176e-94f7-42f8-8a07-85a3f38a9724 service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Received unexpected event network-vif-unplugged-78aa30fc-a52b-4ebd-bb26-1020af74cf2e for instance with vm_state shelved_offloaded and task_state None. [ 833.805986] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 833.806687] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 833.806687] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Deleting the datastore file [datastore1] 4e5dc620-68a1-4e10-a8be-702c4999ca10 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 833.806687] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a7eead32-15e0-483b-837e-fd0ba426a270 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.814344] env[61728]: DEBUG oslo_vmware.api [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Waiting for the task: (returnval){ [ 833.814344] env[61728]: value = "task-464289" [ 833.814344] env[61728]: _type = "Task" [ 833.814344] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.824074] env[61728]: DEBUG oslo_vmware.api [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-464289, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.899680] env[61728]: DEBUG nova.network.neutron [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 833.913046] env[61728]: DEBUG nova.compute.manager [req-15240c86-f8dc-4b1b-9196-e5f746501a8c req-dbb6c48e-d0b4-464e-83c5-2981650e71c3 service nova] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Received event network-vif-plugged-1a5d5fc4-4ae6-467a-8bff-0c0f410c9252 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 833.913309] env[61728]: DEBUG oslo_concurrency.lockutils [req-15240c86-f8dc-4b1b-9196-e5f746501a8c req-dbb6c48e-d0b4-464e-83c5-2981650e71c3 service nova] Acquiring lock "d44718bd-872b-401c-aa11-f10bea4a35d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.913532] env[61728]: DEBUG oslo_concurrency.lockutils [req-15240c86-f8dc-4b1b-9196-e5f746501a8c req-dbb6c48e-d0b4-464e-83c5-2981650e71c3 service nova] Lock "d44718bd-872b-401c-aa11-f10bea4a35d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.914146] env[61728]: DEBUG oslo_concurrency.lockutils [req-15240c86-f8dc-4b1b-9196-e5f746501a8c req-dbb6c48e-d0b4-464e-83c5-2981650e71c3 service nova] Lock "d44718bd-872b-401c-aa11-f10bea4a35d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.914468] env[61728]: DEBUG nova.compute.manager [req-15240c86-f8dc-4b1b-9196-e5f746501a8c req-dbb6c48e-d0b4-464e-83c5-2981650e71c3 service nova] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] No waiting events found dispatching network-vif-plugged-1a5d5fc4-4ae6-467a-8bff-0c0f410c9252 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 833.914720] env[61728]: WARNING nova.compute.manager [req-15240c86-f8dc-4b1b-9196-e5f746501a8c req-dbb6c48e-d0b4-464e-83c5-2981650e71c3 service nova] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Received unexpected event network-vif-plugged-1a5d5fc4-4ae6-467a-8bff-0c0f410c9252 for instance with vm_state building and task_state spawning. [ 834.017392] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464287, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463658} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.017678] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3ef9bf60-c6a1-4b7a-a375-5397fb871850/3ef9bf60-c6a1-4b7a-a375-5397fb871850.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 834.017886] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.018457] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14debb3a-b38f-4a4c-bbde-8cf01dfaf0f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.024947] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 834.024947] env[61728]: value = "task-464290" [ 834.024947] env[61728]: _type = "Task" [ 834.024947] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.032475] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464290, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.080064] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: aec920d1-12fb-4d41-9663-b4c2b7e142d7] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 834.215748] env[61728]: DEBUG nova.network.neutron [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Updating instance_info_cache with network_info: [{"id": "1a5d5fc4-4ae6-467a-8bff-0c0f410c9252", "address": "fa:16:3e:81:20:fc", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a5d5fc4-4a", "ovs_interfaceid": "1a5d5fc4-4ae6-467a-8bff-0c0f410c9252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.222899] env[61728]: DEBUG nova.network.neutron [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Successfully updated port: 02293cc2-3b2d-4b93-b27d-0e61b2780e89 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 834.256656] env[61728]: DEBUG nova.compute.manager [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 834.282904] env[61728]: DEBUG nova.scheduler.client.report [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.290697] env[61728]: DEBUG nova.virt.hardware [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 834.291208] env[61728]: DEBUG nova.virt.hardware [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 834.291208] env[61728]: DEBUG nova.virt.hardware [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 834.291472] env[61728]: DEBUG nova.virt.hardware [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 834.291745] env[61728]: DEBUG nova.virt.hardware [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 834.291974] env[61728]: DEBUG nova.virt.hardware [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 834.292244] env[61728]: DEBUG nova.virt.hardware [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 834.292446] env[61728]: DEBUG nova.virt.hardware [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 834.292658] env[61728]: DEBUG nova.virt.hardware [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 834.292865] env[61728]: DEBUG nova.virt.hardware [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 834.293091] env[61728]: DEBUG nova.virt.hardware [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.294184] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4b8afed-a905-48a3-9ad1-c58fc1ad41dc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.304215] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7593c684-cdcb-443b-b253-3d97c13e44c8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.328015] env[61728]: DEBUG oslo_vmware.api [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Task: {'id': task-464289, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.236473} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.328269] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.328461] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 834.328725] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 834.329189] env[61728]: INFO nova.compute.manager [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Took 1.16 seconds to destroy the instance on the hypervisor. [ 834.329506] env[61728]: DEBUG oslo.service.loopingcall [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.329748] env[61728]: DEBUG nova.compute.manager [-] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.329850] env[61728]: DEBUG nova.network.neutron [-] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 834.537122] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464290, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066454} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.537176] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 834.539304] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98fd6e2-fea4-432d-9505-4adbcdbcddf0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.564034] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 3ef9bf60-c6a1-4b7a-a375-5397fb871850/3ef9bf60-c6a1-4b7a-a375-5397fb871850.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.564815] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b639f97-0e8f-49ce-9b70-d3b787cf5a18 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.583737] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 2c76f45a-b87c-4c3b-9bf9-ba5d67265437] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 834.590573] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 834.590573] env[61728]: value = "task-464291" [ 834.590573] env[61728]: _type = "Task" [ 834.590573] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.604028] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464291, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.719813] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "refresh_cache-d44718bd-872b-401c-aa11-f10bea4a35d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.720184] env[61728]: DEBUG nova.compute.manager [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Instance network_info: |[{"id": "1a5d5fc4-4ae6-467a-8bff-0c0f410c9252", "address": "fa:16:3e:81:20:fc", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a5d5fc4-4a", "ovs_interfaceid": "1a5d5fc4-4ae6-467a-8bff-0c0f410c9252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 834.721435] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:20:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc6d5964-1106-4345-a26d-185dabd4ff0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1a5d5fc4-4ae6-467a-8bff-0c0f410c9252', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 834.731456] env[61728]: DEBUG oslo.service.loopingcall [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.731626] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.732247] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.732497] env[61728]: DEBUG nova.network.neutron [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 834.733789] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 834.734085] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ff20f1b-2bb3-40e6-bdaf-04cad6fd6d6c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.765747] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 834.765747] env[61728]: value = "task-464292" [ 834.765747] env[61728]: _type = "Task" [ 834.765747] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.777446] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464292, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.788200] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.578s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.788789] env[61728]: DEBUG nova.compute.manager [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 834.792118] env[61728]: DEBUG oslo_concurrency.lockutils [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.483s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.792336] env[61728]: DEBUG oslo_concurrency.lockutils [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.794498] env[61728]: DEBUG oslo_concurrency.lockutils [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.515s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.794780] env[61728]: DEBUG oslo_concurrency.lockutils [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.796678] env[61728]: DEBUG oslo_concurrency.lockutils [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 22.344s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.822517] env[61728]: WARNING nova.network.neutron [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] 20ba9b61-3be9-4ad2-a1d0-88c810873bcb already exists in list: networks containing: ['20ba9b61-3be9-4ad2-a1d0-88c810873bcb']. ignoring it [ 834.843390] env[61728]: INFO nova.scheduler.client.report [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Deleted allocations for instance 0fb1192e-99f1-4469-b196-60df7eab8185 [ 834.848996] env[61728]: INFO nova.scheduler.client.report [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Deleted allocations for instance a716cf5a-4482-45cb-96ec-b8c38bc3e742 [ 835.092622] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: c8f1f4ae-ab08-4a03-a3fd-94014509e7e7] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 835.107079] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464291, 'name': ReconfigVM_Task, 'duration_secs': 0.297456} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.107079] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 3ef9bf60-c6a1-4b7a-a375-5397fb871850/3ef9bf60-c6a1-4b7a-a375-5397fb871850.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.107079] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f5f7c60e-846d-400d-9a93-f13a92919327 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.113353] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 835.113353] env[61728]: value = "task-464293" [ 835.113353] env[61728]: _type = "Task" [ 835.113353] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.129180] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464293, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.276995] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464292, 'name': CreateVM_Task, 'duration_secs': 0.330157} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.277614] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 835.277857] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.278046] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.278374] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 835.279042] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ab5ac32-4586-4bf1-91b6-a48a1f538694 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.283781] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 835.283781] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fedaf3-26da-feb5-8f19-96abe33482d7" [ 835.283781] env[61728]: _type = "Task" [ 835.283781] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.292880] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fedaf3-26da-feb5-8f19-96abe33482d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.300777] env[61728]: DEBUG nova.compute.utils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 835.305930] env[61728]: DEBUG nova.compute.manager [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 835.306200] env[61728]: DEBUG nova.network.neutron [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 835.360103] env[61728]: DEBUG oslo_concurrency.lockutils [None req-563e8da4-d118-4061-a985-d0a57a29d8ac tempest-FloatingIPsAssociationTestJSON-691887995 tempest-FloatingIPsAssociationTestJSON-691887995-project-member] Lock "0fb1192e-99f1-4469-b196-60df7eab8185" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.980s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.366976] env[61728]: DEBUG oslo_concurrency.lockutils [None req-319f4716-87b1-445b-8935-a7eaf9aea81b tempest-ServersTestMultiNic-388205169 tempest-ServersTestMultiNic-388205169-project-member] Lock "a716cf5a-4482-45cb-96ec-b8c38bc3e742" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.359s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.371815] env[61728]: DEBUG nova.policy [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88382df814254c3d8330233d4319885a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9fe346266a245dbba86fb9c8cda0a94', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 835.387785] env[61728]: DEBUG nova.network.neutron [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Updating instance_info_cache with network_info: [{"id": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "address": "fa:16:3e:44:e0:e6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e719ade-f3", "ovs_interfaceid": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "02293cc2-3b2d-4b93-b27d-0e61b2780e89", "address": "fa:16:3e:39:43:bf", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02293cc2-3b", "ovs_interfaceid": "02293cc2-3b2d-4b93-b27d-0e61b2780e89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.507547] env[61728]: DEBUG nova.network.neutron [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Successfully updated port: 8904a7fb-a87a-4f80-958e-87748f00b59e {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 835.599702] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: db8e373d-e06c-43f0-a79c-2c1aaf65a01a] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 835.627586] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464293, 'name': Rename_Task, 'duration_secs': 0.135823} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.633029] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 835.633029] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-927a991d-6681-4892-88e4-ad979c34ae81 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.641903] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 835.641903] env[61728]: value = "task-464294" [ 835.641903] env[61728]: _type = "Task" [ 835.641903] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.652221] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464294, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.707960] env[61728]: DEBUG nova.network.neutron [-] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.751954] env[61728]: DEBUG oslo_concurrency.lockutils [None req-935810cb-9d43-4eff-b7bf-b8cb386237f1 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "2976d71f-7322-4640-b734-645dbb70e0c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.764501] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d409ce8-a38b-4727-b8c1-04a8d7309b46 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.773416] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f087c58-2527-4d7e-a701-16fc6a2f4486 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.806404] env[61728]: DEBUG nova.compute.manager [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 835.814492] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f600eb-f359-4501-9769-ecba7dbaaa15 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.826117] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3eeac86-827f-4c1a-b231-e4fe4f59c127 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.828903] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52fedaf3-26da-feb5-8f19-96abe33482d7, 'name': SearchDatastore_Task, 'duration_secs': 0.014155} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.828903] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.829951] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.829951] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.829951] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.829951] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.830299] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-377e063f-f116-49ff-be62-ee057d7496a8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.840637] env[61728]: DEBUG nova.compute.provider_tree [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.848950] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.849216] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 835.849981] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41692286-90d9-4429-b998-921519e10c4b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.857378] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 835.857378] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5220a725-8b87-a633-f320-037a9e667858" [ 835.857378] env[61728]: _type = "Task" [ 835.857378] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.870116] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5220a725-8b87-a633-f320-037a9e667858, 'name': SearchDatastore_Task, 'duration_secs': 0.008663} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.870918] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb93094f-5383-40f3-9c83-f2545e4e0461 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.876230] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 835.876230] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522c6cd2-e6cc-c8a0-d5ee-3038092706fc" [ 835.876230] env[61728]: _type = "Task" [ 835.876230] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.884658] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522c6cd2-e6cc-c8a0-d5ee-3038092706fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.890330] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.890984] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.891162] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.895064] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317e1ef3-c852-477d-a8ee-71f3a7a64631 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.910314] env[61728]: DEBUG nova.virt.hardware [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.910592] env[61728]: DEBUG nova.virt.hardware [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.910760] env[61728]: DEBUG nova.virt.hardware [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.910952] env[61728]: DEBUG nova.virt.hardware [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.911119] env[61728]: DEBUG nova.virt.hardware [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.911278] env[61728]: DEBUG nova.virt.hardware [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.911492] env[61728]: DEBUG nova.virt.hardware [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.911667] env[61728]: DEBUG nova.virt.hardware [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.911851] env[61728]: DEBUG nova.virt.hardware [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.912026] env[61728]: DEBUG nova.virt.hardware [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.912205] env[61728]: DEBUG nova.virt.hardware [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.919062] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Reconfiguring VM to attach interface {{(pid=61728) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 835.919426] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60a76b7d-0c74-47f3-b628-4cbbc19fb6d9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.937704] env[61728]: DEBUG oslo_vmware.api [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 835.937704] env[61728]: value = "task-464295" [ 835.937704] env[61728]: _type = "Task" [ 835.937704] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.946496] env[61728]: DEBUG oslo_vmware.api [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464295, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.011483] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "refresh_cache-b444a6cf-3c98-4308-afc1-6e760d30082d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.011794] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquired lock "refresh_cache-b444a6cf-3c98-4308-afc1-6e760d30082d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.011964] env[61728]: DEBUG nova.network.neutron [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 836.103380] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: be014f77-53ca-42a5-9255-2ba950cca314] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 836.151304] env[61728]: DEBUG oslo_vmware.api [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464294, 'name': PowerOnVM_Task, 'duration_secs': 0.480336} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.151536] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 836.151765] env[61728]: INFO nova.compute.manager [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Took 7.11 seconds to spawn the instance on the hypervisor. [ 836.152013] env[61728]: DEBUG nova.compute.manager [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 836.152844] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fa53078-e87d-420a-ae7e-8198e5434098 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.211530] env[61728]: INFO nova.compute.manager [-] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Took 1.88 seconds to deallocate network for instance. [ 836.240375] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.240717] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.240979] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.241235] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.241452] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.246685] env[61728]: INFO nova.compute.manager [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Terminating instance [ 836.249018] env[61728]: DEBUG nova.compute.manager [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 836.249274] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 836.250284] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f1e9c0-67cc-4cb8-960e-5fa572be594d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.260863] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 836.261743] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76b6fa9d-6604-4aba-a190-bcbdfd8e16ec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.269473] env[61728]: DEBUG oslo_vmware.api [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 836.269473] env[61728]: value = "task-464296" [ 836.269473] env[61728]: _type = "Task" [ 836.269473] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.277379] env[61728]: DEBUG oslo_vmware.api [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464296, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.346282] env[61728]: DEBUG nova.scheduler.client.report [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.353328] env[61728]: DEBUG nova.network.neutron [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Successfully created port: 6969ab65-3bd4-4500-9908-9fe9949e1bab {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 836.386593] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522c6cd2-e6cc-c8a0-d5ee-3038092706fc, 'name': SearchDatastore_Task, 'duration_secs': 0.008837} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.386700] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.390028] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] d44718bd-872b-401c-aa11-f10bea4a35d8/d44718bd-872b-401c-aa11-f10bea4a35d8.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 836.390028] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d83aa5a6-6cf9-445c-b6de-f716034103d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.396325] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 836.396325] env[61728]: value = "task-464297" [ 836.396325] env[61728]: _type = "Task" [ 836.396325] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.404673] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464297, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.448124] env[61728]: DEBUG oslo_vmware.api [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464295, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.488628] env[61728]: DEBUG nova.compute.manager [req-2fd5634f-3d49-45a1-805b-861d94324a93 req-e7d7119b-4e6d-40f4-9eeb-83a56e700c3a service nova] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Received event network-vif-plugged-8904a7fb-a87a-4f80-958e-87748f00b59e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 836.489021] env[61728]: DEBUG oslo_concurrency.lockutils [req-2fd5634f-3d49-45a1-805b-861d94324a93 req-e7d7119b-4e6d-40f4-9eeb-83a56e700c3a service nova] Acquiring lock "b444a6cf-3c98-4308-afc1-6e760d30082d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.489142] env[61728]: DEBUG oslo_concurrency.lockutils [req-2fd5634f-3d49-45a1-805b-861d94324a93 req-e7d7119b-4e6d-40f4-9eeb-83a56e700c3a service nova] Lock "b444a6cf-3c98-4308-afc1-6e760d30082d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.489322] env[61728]: DEBUG oslo_concurrency.lockutils [req-2fd5634f-3d49-45a1-805b-861d94324a93 req-e7d7119b-4e6d-40f4-9eeb-83a56e700c3a service nova] Lock "b444a6cf-3c98-4308-afc1-6e760d30082d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.489726] env[61728]: DEBUG nova.compute.manager [req-2fd5634f-3d49-45a1-805b-861d94324a93 req-e7d7119b-4e6d-40f4-9eeb-83a56e700c3a service nova] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] No waiting events found dispatching network-vif-plugged-8904a7fb-a87a-4f80-958e-87748f00b59e {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 836.489872] env[61728]: WARNING nova.compute.manager [req-2fd5634f-3d49-45a1-805b-861d94324a93 req-e7d7119b-4e6d-40f4-9eeb-83a56e700c3a service nova] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Received unexpected event network-vif-plugged-8904a7fb-a87a-4f80-958e-87748f00b59e for instance with vm_state building and task_state spawning. [ 836.559309] env[61728]: DEBUG nova.network.neutron [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 836.607193] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: a0831461-ece1-43ee-92f6-34d7d4e673e2] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 836.675063] env[61728]: INFO nova.compute.manager [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Took 65.03 seconds to build instance. [ 836.721386] env[61728]: DEBUG oslo_concurrency.lockutils [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.782233] env[61728]: DEBUG oslo_vmware.api [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464296, 'name': PowerOffVM_Task, 'duration_secs': 0.265519} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.783467] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 836.783467] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 836.783467] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b7b4979d-cca4-4589-b95f-77b96117221f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.802687] env[61728]: DEBUG nova.network.neutron [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Updating instance_info_cache with network_info: [{"id": "8904a7fb-a87a-4f80-958e-87748f00b59e", "address": "fa:16:3e:10:db:9d", "network": {"id": "22e73f4c-99d9-4e65-8a72-f6ea53d88eaf", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2136421023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "29dfac4a253f43be811df957e0ee5f96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8904a7fb-a8", "ovs_interfaceid": "8904a7fb-a87a-4f80-958e-87748f00b59e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.816335] env[61728]: DEBUG nova.compute.manager [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 836.850946] env[61728]: DEBUG nova.virt.hardware [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.851384] env[61728]: DEBUG nova.virt.hardware [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.851559] env[61728]: DEBUG nova.virt.hardware [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.851878] env[61728]: DEBUG nova.virt.hardware [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.851967] env[61728]: DEBUG nova.virt.hardware [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.852091] env[61728]: DEBUG nova.virt.hardware [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.852529] env[61728]: DEBUG nova.virt.hardware [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.852831] env[61728]: DEBUG nova.virt.hardware [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.853255] env[61728]: DEBUG nova.virt.hardware [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.853488] env[61728]: DEBUG nova.virt.hardware [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.853686] env[61728]: DEBUG nova.virt.hardware [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.854951] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90333b64-f8a7-41a8-b290-f9778e96865c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.869854] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490cc2c0-fcba-4448-b016-5112bb5a7717 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.878817] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 836.879137] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 836.879433] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Deleting the datastore file [datastore1] bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 836.880159] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14743177-0d44-4e00-8b83-b9040b2aa715 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.903738] env[61728]: DEBUG oslo_vmware.api [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for the task: (returnval){ [ 836.903738] env[61728]: value = "task-464299" [ 836.903738] env[61728]: _type = "Task" [ 836.903738] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.913835] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464297, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.919707] env[61728]: DEBUG oslo_vmware.api [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464299, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.948569] env[61728]: DEBUG oslo_vmware.api [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464295, 'name': ReconfigVM_Task, 'duration_secs': 0.670569} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.949453] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.949824] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Reconfigured VM to attach interface {{(pid=61728) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 837.110320] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 8c4b7227-1a01-4400-beb8-2df27e17e329] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 837.177051] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa932093-9d79-4adf-a19c-9f7a8d97d63c tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "3ef9bf60-c6a1-4b7a-a375-5397fb871850" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.581s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.306383] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Releasing lock "refresh_cache-b444a6cf-3c98-4308-afc1-6e760d30082d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.306935] env[61728]: DEBUG nova.compute.manager [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Instance network_info: |[{"id": "8904a7fb-a87a-4f80-958e-87748f00b59e", "address": "fa:16:3e:10:db:9d", "network": {"id": "22e73f4c-99d9-4e65-8a72-f6ea53d88eaf", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2136421023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "29dfac4a253f43be811df957e0ee5f96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8904a7fb-a8", "ovs_interfaceid": "8904a7fb-a87a-4f80-958e-87748f00b59e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 837.307475] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:db:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6685c85e-be1e-4b7b-a6cc-3e50e59b6567', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8904a7fb-a87a-4f80-958e-87748f00b59e', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.317804] env[61728]: DEBUG oslo.service.loopingcall [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.318697] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 837.319202] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-99c1e063-d688-48dc-ba3d-3c3f4f3e1b99 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.346390] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.346390] env[61728]: value = "task-464300" [ 837.346390] env[61728]: _type = "Task" [ 837.346390] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.355578] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464300, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.365283] env[61728]: DEBUG oslo_concurrency.lockutils [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.568s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.369803] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.451s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.370539] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.373751] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.393s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.375032] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.376296] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.973s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.376554] env[61728]: DEBUG nova.objects.instance [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lazy-loading 'resources' on Instance uuid dc8a587f-3dc8-43b3-b986-2096e3c4e1f1 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 837.423140] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464297, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.563377} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.426114] env[61728]: INFO nova.scheduler.client.report [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleted allocations for instance a3e9af38-5115-4337-a20e-5d6c6ec655fe [ 837.429623] env[61728]: INFO nova.scheduler.client.report [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Deleted allocations for instance b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4 [ 837.433289] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] d44718bd-872b-401c-aa11-f10bea4a35d8/d44718bd-872b-401c-aa11-f10bea4a35d8.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 837.433289] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.433289] env[61728]: DEBUG oslo_vmware.api [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Task: {'id': task-464299, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181913} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.440270] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-96937d55-c119-4cd4-8b78-1d53eee3e0ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.446194] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 837.446194] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 837.446194] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 837.446194] env[61728]: INFO nova.compute.manager [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Took 1.20 seconds to destroy the instance on the hypervisor. [ 837.446194] env[61728]: DEBUG oslo.service.loopingcall [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.446194] env[61728]: DEBUG nova.compute.manager [-] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 837.446194] env[61728]: DEBUG nova.network.neutron [-] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 837.452745] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 837.452745] env[61728]: value = "task-464301" [ 837.452745] env[61728]: _type = "Task" [ 837.452745] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.458624] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ee5691c-816c-4870-953d-b39293df819f tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-f28348d2-c062-497a-b374-521df51054ee-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.530s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.471933] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464301, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.613637] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 620bb472-c36e-4c56-acdf-42e02614856b] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 837.860416] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464300, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.951253] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c8544d76-2d00-4a25-9137-a46b374e89e9 tempest-ServerRescueTestJSONUnderV235-1512673932 tempest-ServerRescueTestJSONUnderV235-1512673932-project-member] Lock "b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.698s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.955651] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7eb06215-5599-4a06-aa7a-34d991fbcb08 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "a3e9af38-5115-4337-a20e-5d6c6ec655fe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.221s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.962646] env[61728]: INFO nova.scheduler.client.report [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleted allocation for migration adc30b71-8803-41e0-a72a-09cb8bee068c [ 837.972218] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464301, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090771} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.973374] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 837.976735] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc8304aa-f712-4b36-9493-4951e2858fd6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.001963] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] d44718bd-872b-401c-aa11-f10bea4a35d8/d44718bd-872b-401c-aa11-f10bea4a35d8.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.005620] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-429d26ec-b670-4234-a9bc-d2cbe5e23f2d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.028922] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 838.028922] env[61728]: value = "task-464302" [ 838.028922] env[61728]: _type = "Task" [ 838.028922] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.043157] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464302, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.119872] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 3c4f7d53-6ba6-4e2c-b324-e651640b304b] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 838.149836] env[61728]: DEBUG nova.compute.manager [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Received event network-changed-78aa30fc-a52b-4ebd-bb26-1020af74cf2e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 838.149836] env[61728]: DEBUG nova.compute.manager [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Refreshing instance network info cache due to event network-changed-78aa30fc-a52b-4ebd-bb26-1020af74cf2e. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 838.149836] env[61728]: DEBUG oslo_concurrency.lockutils [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] Acquiring lock "refresh_cache-2976d71f-7322-4640-b734-645dbb70e0c3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.149836] env[61728]: DEBUG oslo_concurrency.lockutils [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] Acquired lock "refresh_cache-2976d71f-7322-4640-b734-645dbb70e0c3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.149836] env[61728]: DEBUG nova.network.neutron [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Refreshing network info cache for port 78aa30fc-a52b-4ebd-bb26-1020af74cf2e {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 838.195808] env[61728]: DEBUG nova.compute.manager [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Received event network-changed-1a5d5fc4-4ae6-467a-8bff-0c0f410c9252 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 838.196554] env[61728]: DEBUG nova.compute.manager [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Refreshing instance network info cache due to event network-changed-1a5d5fc4-4ae6-467a-8bff-0c0f410c9252. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 838.196851] env[61728]: DEBUG oslo_concurrency.lockutils [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] Acquiring lock "refresh_cache-d44718bd-872b-401c-aa11-f10bea4a35d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.197111] env[61728]: DEBUG oslo_concurrency.lockutils [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] Acquired lock "refresh_cache-d44718bd-872b-401c-aa11-f10bea4a35d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.197817] env[61728]: DEBUG nova.network.neutron [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Refreshing network info cache for port 1a5d5fc4-4ae6-467a-8bff-0c0f410c9252 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 838.343637] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc14ba8b-9087-48d5-bc7b-62bfac1c7e74 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.354234] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd3d7436-e572-4bad-9227-08b26b0b27c5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.360363] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464300, 'name': CreateVM_Task, 'duration_secs': 0.798167} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.360915] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 838.362399] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.362668] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.363047] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.363740] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcd91b8e-2fbe-447b-8f0a-cfc90cacf350 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.391287] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e0208b8-c8b4-4b85-bd9a-11148cfe98fa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.396140] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 838.396140] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d5ee4d-571b-f00d-07c6-116d820a64de" [ 838.396140] env[61728]: _type = "Task" [ 838.396140] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.403710] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741a2b15-3083-415f-ac2b-b0df866f6767 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.413092] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d5ee4d-571b-f00d-07c6-116d820a64de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.424447] env[61728]: DEBUG nova.compute.provider_tree [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.469092] env[61728]: DEBUG oslo_concurrency.lockutils [None req-44a5becc-1b57-429a-b527-d342dd355387 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "f2692e22-6c14-4bb3-a080-607f0731105d" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 29.393s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.521492] env[61728]: DEBUG nova.network.neutron [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Successfully updated port: 6969ab65-3bd4-4500-9908-9fe9949e1bab {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 838.547360] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464302, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.628351] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: e56a59a6-fc0f-45ca-810a-76b18bae173a] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 838.909154] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d5ee4d-571b-f00d-07c6-116d820a64de, 'name': SearchDatastore_Task, 'duration_secs': 0.017452} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.914698] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.915025] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.915386] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.915698] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.916020] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.919164] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-266de174-c825-4726-936d-3a908be691b3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.929344] env[61728]: DEBUG nova.scheduler.client.report [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.937106] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.939701] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 838.941391] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11259246-3045-4a12-a732-94580efac306 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.948882] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 838.948882] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52314f7b-33b6-c983-7043-501f22aada4e" [ 838.948882] env[61728]: _type = "Task" [ 838.948882] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.958669] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52314f7b-33b6-c983-7043-501f22aada4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.027829] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Acquiring lock "refresh_cache-174607a1-9bc8-4e07-8993-7f0bb0f308e9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.027972] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Acquired lock "refresh_cache-174607a1-9bc8-4e07-8993-7f0bb0f308e9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.029394] env[61728]: DEBUG nova.network.neutron [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 839.041462] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464302, 'name': ReconfigVM_Task, 'duration_secs': 0.901107} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.043065] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Reconfigured VM instance instance-00000045 to attach disk [datastore1] d44718bd-872b-401c-aa11-f10bea4a35d8/d44718bd-872b-401c-aa11-f10bea4a35d8.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.044291] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad2ebabf-a70c-479e-b1ec-a80ced4db171 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.052379] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 839.052379] env[61728]: value = "task-464303" [ 839.052379] env[61728]: _type = "Task" [ 839.052379] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.061959] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464303, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.132280] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: adb58292-62cd-4142-ad66-db014bcbdc49] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 839.193655] env[61728]: DEBUG nova.network.neutron [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Updated VIF entry in instance network info cache for port 78aa30fc-a52b-4ebd-bb26-1020af74cf2e. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 839.194121] env[61728]: DEBUG nova.network.neutron [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Updating instance_info_cache with network_info: [{"id": "78aa30fc-a52b-4ebd-bb26-1020af74cf2e", "address": "fa:16:3e:16:c0:48", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": null, "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap78aa30fc-a5", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.229674] env[61728]: DEBUG nova.network.neutron [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Updated VIF entry in instance network info cache for port 1a5d5fc4-4ae6-467a-8bff-0c0f410c9252. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 839.230075] env[61728]: DEBUG nova.network.neutron [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Updating instance_info_cache with network_info: [{"id": "1a5d5fc4-4ae6-467a-8bff-0c0f410c9252", "address": "fa:16:3e:81:20:fc", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a5d5fc4-4a", "ovs_interfaceid": "1a5d5fc4-4ae6-467a-8bff-0c0f410c9252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.238069] env[61728]: DEBUG nova.network.neutron [-] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.442489] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.066s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.445703] env[61728]: DEBUG oslo_concurrency.lockutils [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.908s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.445973] env[61728]: DEBUG nova.objects.instance [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Lazy-loading 'resources' on Instance uuid a5c7e3ff-6829-4528-9279-e865a0eb8512 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 839.468814] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52314f7b-33b6-c983-7043-501f22aada4e, 'name': SearchDatastore_Task, 'duration_secs': 0.012028} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.470262] env[61728]: INFO nova.scheduler.client.report [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Deleted allocations for instance dc8a587f-3dc8-43b3-b986-2096e3c4e1f1 [ 839.478854] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9a77447-efa8-4dc2-9268-dfff3ba5e919 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.486311] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 839.486311] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f4fa60-70a3-ecdd-3f62-d25460fa766f" [ 839.486311] env[61728]: _type = "Task" [ 839.486311] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.496811] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f4fa60-70a3-ecdd-3f62-d25460fa766f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.563369] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464303, 'name': Rename_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.598933] env[61728]: DEBUG nova.network.neutron [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 839.636736] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: b86bf217-d1f8-4d37-aa87-3a2d27d70420] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 839.696457] env[61728]: DEBUG oslo_concurrency.lockutils [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] Releasing lock "refresh_cache-2976d71f-7322-4640-b734-645dbb70e0c3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.697046] env[61728]: DEBUG nova.compute.manager [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Received event network-changed-8904a7fb-a87a-4f80-958e-87748f00b59e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.697292] env[61728]: DEBUG nova.compute.manager [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Refreshing instance network info cache due to event network-changed-8904a7fb-a87a-4f80-958e-87748f00b59e. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 839.697528] env[61728]: DEBUG oslo_concurrency.lockutils [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] Acquiring lock "refresh_cache-b444a6cf-3c98-4308-afc1-6e760d30082d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.697678] env[61728]: DEBUG oslo_concurrency.lockutils [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] Acquired lock "refresh_cache-b444a6cf-3c98-4308-afc1-6e760d30082d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.697854] env[61728]: DEBUG nova.network.neutron [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Refreshing network info cache for port 8904a7fb-a87a-4f80-958e-87748f00b59e {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 839.735234] env[61728]: DEBUG oslo_concurrency.lockutils [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] Releasing lock "refresh_cache-d44718bd-872b-401c-aa11-f10bea4a35d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.735234] env[61728]: DEBUG nova.compute.manager [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Received event network-vif-plugged-02293cc2-3b2d-4b93-b27d-0e61b2780e89 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.735234] env[61728]: DEBUG oslo_concurrency.lockutils [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] Acquiring lock "f28348d2-c062-497a-b374-521df51054ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.735234] env[61728]: DEBUG oslo_concurrency.lockutils [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] Lock "f28348d2-c062-497a-b374-521df51054ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.735234] env[61728]: DEBUG oslo_concurrency.lockutils [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] Lock "f28348d2-c062-497a-b374-521df51054ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.735234] env[61728]: DEBUG nova.compute.manager [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] No waiting events found dispatching network-vif-plugged-02293cc2-3b2d-4b93-b27d-0e61b2780e89 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 839.735234] env[61728]: WARNING nova.compute.manager [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Received unexpected event network-vif-plugged-02293cc2-3b2d-4b93-b27d-0e61b2780e89 for instance with vm_state active and task_state None. [ 839.735234] env[61728]: DEBUG nova.compute.manager [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Received event network-changed-02293cc2-3b2d-4b93-b27d-0e61b2780e89 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.735234] env[61728]: DEBUG nova.compute.manager [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Refreshing instance network info cache due to event network-changed-02293cc2-3b2d-4b93-b27d-0e61b2780e89. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 839.735234] env[61728]: DEBUG oslo_concurrency.lockutils [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] Acquiring lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.735234] env[61728]: DEBUG oslo_concurrency.lockutils [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] Acquired lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.735234] env[61728]: DEBUG nova.network.neutron [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Refreshing network info cache for port 02293cc2-3b2d-4b93-b27d-0e61b2780e89 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 839.743153] env[61728]: INFO nova.compute.manager [-] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Took 2.30 seconds to deallocate network for instance. [ 839.862436] env[61728]: DEBUG nova.network.neutron [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Updating instance_info_cache with network_info: [{"id": "6969ab65-3bd4-4500-9908-9fe9949e1bab", "address": "fa:16:3e:49:18:80", "network": {"id": "d24b80d4-9eda-45b0-b1bb-cd522a1260fc", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1872056846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9fe346266a245dbba86fb9c8cda0a94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3", "external-id": "nsx-vlan-transportzone-263", "segmentation_id": 263, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6969ab65-3b", "ovs_interfaceid": "6969ab65-3bd4-4500-9908-9fe9949e1bab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.990490] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a44611e3-0df9-42b8-8328-95836c27c204 tempest-ServerShowV247Test-644229886 tempest-ServerShowV247Test-644229886-project-member] Lock "dc8a587f-3dc8-43b3-b986-2096e3c4e1f1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.355s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.002063] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f4fa60-70a3-ecdd-3f62-d25460fa766f, 'name': SearchDatastore_Task, 'duration_secs': 0.010672} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.004720] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.005590] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b444a6cf-3c98-4308-afc1-6e760d30082d/b444a6cf-3c98-4308-afc1-6e760d30082d.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 840.005899] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-409b0c60-c9b7-44eb-b1d0-badee46697c3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.018122] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 840.018122] env[61728]: value = "task-464304" [ 840.018122] env[61728]: _type = "Task" [ 840.018122] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.035200] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464304, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.071364] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464303, 'name': Rename_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.133691] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "771341ed-8b8e-470c-9686-82650f5271b4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.134098] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "771341ed-8b8e-470c-9686-82650f5271b4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.134272] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "771341ed-8b8e-470c-9686-82650f5271b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.134482] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "771341ed-8b8e-470c-9686-82650f5271b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.137557] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "771341ed-8b8e-470c-9686-82650f5271b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.137557] env[61728]: INFO nova.compute.manager [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Terminating instance [ 840.139949] env[61728]: DEBUG nova.compute.manager [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 840.139949] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 840.142114] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95af5cbd-96e7-47b7-b2d7-1f72ef6dbd02 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.148311] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 4247e356-ef10-41b3-bb77-106103134d59] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 840.155767] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 840.156058] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc1b5232-bce8-4eed-a0f5-0059f686e88f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.164144] env[61728]: DEBUG oslo_vmware.api [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 840.164144] env[61728]: value = "task-464305" [ 840.164144] env[61728]: _type = "Task" [ 840.164144] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.177017] env[61728]: DEBUG oslo_vmware.api [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-464305, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.253098] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.365910] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Releasing lock "refresh_cache-174607a1-9bc8-4e07-8993-7f0bb0f308e9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.366053] env[61728]: DEBUG nova.compute.manager [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Instance network_info: |[{"id": "6969ab65-3bd4-4500-9908-9fe9949e1bab", "address": "fa:16:3e:49:18:80", "network": {"id": "d24b80d4-9eda-45b0-b1bb-cd522a1260fc", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1872056846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9fe346266a245dbba86fb9c8cda0a94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3", "external-id": "nsx-vlan-transportzone-263", "segmentation_id": 263, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6969ab65-3b", "ovs_interfaceid": "6969ab65-3bd4-4500-9908-9fe9949e1bab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 840.366487] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:18:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6969ab65-3bd4-4500-9908-9fe9949e1bab', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 840.376681] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Creating folder: Project (f9fe346266a245dbba86fb9c8cda0a94). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 840.379743] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e5e590d6-d6c4-4001-ab9a-d1129c182d08 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.391334] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Created folder: Project (f9fe346266a245dbba86fb9c8cda0a94) in parent group-v121913. [ 840.391334] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Creating folder: Instances. Parent ref: group-v122120. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 840.391334] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-490553ae-3f16-4477-93fb-18cbfa5f9123 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.403337] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Created folder: Instances in parent group-v122120. [ 840.403681] env[61728]: DEBUG oslo.service.loopingcall [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.403681] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 840.403917] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e9a36a39-4389-44bb-a9ba-58ef4f0ad37f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.431345] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 840.431345] env[61728]: value = "task-464308" [ 840.431345] env[61728]: _type = "Task" [ 840.431345] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.439286] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "3ef9bf60-c6a1-4b7a-a375-5397fb871850" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.439452] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "3ef9bf60-c6a1-4b7a-a375-5397fb871850" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.439701] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "3ef9bf60-c6a1-4b7a-a375-5397fb871850-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.440261] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "3ef9bf60-c6a1-4b7a-a375-5397fb871850-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.440261] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "3ef9bf60-c6a1-4b7a-a375-5397fb871850-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.444558] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956e849d-3a39-4513-801b-29f7a1cc9e2d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.447311] env[61728]: INFO nova.compute.manager [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Terminating instance [ 840.455089] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464308, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.455935] env[61728]: DEBUG nova.compute.manager [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 840.456205] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 840.457848] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83454fc9-bec0-4bad-b3a1-efebdd2e8c01 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.465469] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec79b72f-a77c-4a6d-8f2b-53cef11431ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.471328] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 840.472101] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46725d42-6966-40d2-890a-b60a028d3e17 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.510039] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e20c3ed-5168-461f-b467-33addea21699 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.513070] env[61728]: DEBUG oslo_vmware.api [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 840.513070] env[61728]: value = "task-464309" [ 840.513070] env[61728]: _type = "Task" [ 840.513070] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.521377] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-234c2fa7-70eb-472c-85b4-8ae4ae407beb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.528728] env[61728]: DEBUG oslo_vmware.api [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464309, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.542355] env[61728]: DEBUG nova.compute.provider_tree [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.547106] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464304, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509601} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.547615] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b444a6cf-3c98-4308-afc1-6e760d30082d/b444a6cf-3c98-4308-afc1-6e760d30082d.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 840.547834] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.548144] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bda88523-2ca6-446e-9451-cd7e98a83a89 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.556448] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 840.556448] env[61728]: value = "task-464310" [ 840.556448] env[61728]: _type = "Task" [ 840.556448] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.576171] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464310, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.581397] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464303, 'name': Rename_Task, 'duration_secs': 1.230566} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.581397] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 840.581397] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2b8e5ed5-70ea-4c7e-a92a-b71a67746a1c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.588378] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 840.588378] env[61728]: value = "task-464311" [ 840.588378] env[61728]: _type = "Task" [ 840.588378] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.596610] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464311, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.652745] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: ee2bf649-3ecc-4f44-a6a7-f4a47bcb6618] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 840.677186] env[61728]: DEBUG oslo_vmware.api [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-464305, 'name': PowerOffVM_Task, 'duration_secs': 0.506799} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.677533] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 840.677707] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 840.677973] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-89716f7c-bb03-4850-b653-a087c204dd84 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.719332] env[61728]: DEBUG nova.network.neutron [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Updated VIF entry in instance network info cache for port 8904a7fb-a87a-4f80-958e-87748f00b59e. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 840.719755] env[61728]: DEBUG nova.network.neutron [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Updating instance_info_cache with network_info: [{"id": "8904a7fb-a87a-4f80-958e-87748f00b59e", "address": "fa:16:3e:10:db:9d", "network": {"id": "22e73f4c-99d9-4e65-8a72-f6ea53d88eaf", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2136421023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "29dfac4a253f43be811df957e0ee5f96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8904a7fb-a8", "ovs_interfaceid": "8904a7fb-a87a-4f80-958e-87748f00b59e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.943424] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464308, 'name': CreateVM_Task, 'duration_secs': 0.469128} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.943626] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 840.944361] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.944543] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.944869] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 840.945172] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ef52e3b-1502-4db3-936c-d88b56180889 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.950586] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Waiting for the task: (returnval){ [ 840.950586] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526ddad4-45e3-1b71-569e-1223c514288d" [ 840.950586] env[61728]: _type = "Task" [ 840.950586] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.959571] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526ddad4-45e3-1b71-569e-1223c514288d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.991182] env[61728]: DEBUG nova.network.neutron [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Updated VIF entry in instance network info cache for port 02293cc2-3b2d-4b93-b27d-0e61b2780e89. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 840.991182] env[61728]: DEBUG nova.network.neutron [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Updating instance_info_cache with network_info: [{"id": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "address": "fa:16:3e:44:e0:e6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e719ade-f3", "ovs_interfaceid": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "02293cc2-3b2d-4b93-b27d-0e61b2780e89", "address": "fa:16:3e:39:43:bf", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02293cc2-3b", "ovs_interfaceid": "02293cc2-3b2d-4b93-b27d-0e61b2780e89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.028293] env[61728]: DEBUG oslo_vmware.api [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464309, 'name': PowerOffVM_Task, 'duration_secs': 0.22765} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.028293] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 841.028293] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 841.028293] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5e6c3911-63f4-44c1-9ae0-75fa53ff9bca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.040888] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 841.040888] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 841.042350] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleting the datastore file [datastore1] 771341ed-8b8e-470c-9686-82650f5271b4 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.042350] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e045d913-f8d6-4731-9867-b910e992b3eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.050853] env[61728]: DEBUG nova.scheduler.client.report [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 841.059028] env[61728]: DEBUG oslo_vmware.api [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for the task: (returnval){ [ 841.059028] env[61728]: value = "task-464314" [ 841.059028] env[61728]: _type = "Task" [ 841.059028] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.068288] env[61728]: DEBUG oslo_vmware.api [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-464314, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.073318] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464310, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084628} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.074268] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.077105] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee57497-2a2f-47dc-926b-399b71f0391d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.107462] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] b444a6cf-3c98-4308-afc1-6e760d30082d/b444a6cf-3c98-4308-afc1-6e760d30082d.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.114071] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc293fa0-8781-4439-811b-602736b067e8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.128482] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 841.128726] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 841.128910] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Deleting the datastore file [datastore1] 3ef9bf60-c6a1-4b7a-a375-5397fb871850 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.130141] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2656e6a9-34fc-49c7-96b1-afc4adcdf5de {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.139370] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464311, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.142073] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 841.142073] env[61728]: value = "task-464315" [ 841.142073] env[61728]: _type = "Task" [ 841.142073] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.142253] env[61728]: DEBUG oslo_vmware.api [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for the task: (returnval){ [ 841.142253] env[61728]: value = "task-464316" [ 841.142253] env[61728]: _type = "Task" [ 841.142253] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.156975] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 26220c0c-f535-4315-b3bb-2e5ac4d2286e] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 841.158864] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464315, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.162168] env[61728]: DEBUG oslo_vmware.api [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.224028] env[61728]: DEBUG oslo_concurrency.lockutils [req-5e6ca8a0-6379-4ac5-b20a-951be35a8a1d req-6f02b83f-63f2-4ee0-85c4-3287c0357d9a service nova] Releasing lock "refresh_cache-b444a6cf-3c98-4308-afc1-6e760d30082d" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.370958] env[61728]: DEBUG nova.compute.manager [req-dc4f2b55-0ea6-458c-a246-cbfddaac05cc req-20929b31-b3ce-4660-b0d7-e4820305729a service nova] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Received event network-vif-deleted-8fcb3f51-74e2-4548-8b86-bc5bc9156f40 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.465712] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526ddad4-45e3-1b71-569e-1223c514288d, 'name': SearchDatastore_Task, 'duration_secs': 0.013357} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.465712] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.466519] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 841.466862] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.467150] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.467565] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 841.468707] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff8b19c7-d751-4906-a0ab-1d92812916a2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.483462] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 841.483786] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 841.484684] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4e649dd-1865-48b5-9679-fbbbda42185f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.491354] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Waiting for the task: (returnval){ [ 841.491354] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c669b3-716d-0011-e53f-a4043be9094c" [ 841.491354] env[61728]: _type = "Task" [ 841.491354] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.492065] env[61728]: DEBUG oslo_concurrency.lockutils [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] Releasing lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.492330] env[61728]: DEBUG nova.compute.manager [req-83335380-6838-47f4-b5da-3ad64bbc6b1b req-ea49d6e0-95d5-4cf7-a39a-c34bb998b2ff service nova] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Received event network-vif-deleted-6408b220-22b3-4b0d-961e-9a7a9ae42648 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.503166] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c669b3-716d-0011-e53f-a4043be9094c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.535359] env[61728]: DEBUG nova.compute.manager [req-5c635c12-afae-4523-8a08-31688891be10 req-fd92efab-cea7-49c8-808f-a991d181e019 service nova] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Received event network-vif-plugged-6969ab65-3bd4-4500-9908-9fe9949e1bab {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.535359] env[61728]: DEBUG oslo_concurrency.lockutils [req-5c635c12-afae-4523-8a08-31688891be10 req-fd92efab-cea7-49c8-808f-a991d181e019 service nova] Acquiring lock "174607a1-9bc8-4e07-8993-7f0bb0f308e9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.535359] env[61728]: DEBUG oslo_concurrency.lockutils [req-5c635c12-afae-4523-8a08-31688891be10 req-fd92efab-cea7-49c8-808f-a991d181e019 service nova] Lock "174607a1-9bc8-4e07-8993-7f0bb0f308e9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.535359] env[61728]: DEBUG oslo_concurrency.lockutils [req-5c635c12-afae-4523-8a08-31688891be10 req-fd92efab-cea7-49c8-808f-a991d181e019 service nova] Lock "174607a1-9bc8-4e07-8993-7f0bb0f308e9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.535359] env[61728]: DEBUG nova.compute.manager [req-5c635c12-afae-4523-8a08-31688891be10 req-fd92efab-cea7-49c8-808f-a991d181e019 service nova] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] No waiting events found dispatching network-vif-plugged-6969ab65-3bd4-4500-9908-9fe9949e1bab {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 841.535359] env[61728]: WARNING nova.compute.manager [req-5c635c12-afae-4523-8a08-31688891be10 req-fd92efab-cea7-49c8-808f-a991d181e019 service nova] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Received unexpected event network-vif-plugged-6969ab65-3bd4-4500-9908-9fe9949e1bab for instance with vm_state building and task_state spawning. [ 841.535359] env[61728]: DEBUG nova.compute.manager [req-5c635c12-afae-4523-8a08-31688891be10 req-fd92efab-cea7-49c8-808f-a991d181e019 service nova] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Received event network-changed-6969ab65-3bd4-4500-9908-9fe9949e1bab {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.535359] env[61728]: DEBUG nova.compute.manager [req-5c635c12-afae-4523-8a08-31688891be10 req-fd92efab-cea7-49c8-808f-a991d181e019 service nova] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Refreshing instance network info cache due to event network-changed-6969ab65-3bd4-4500-9908-9fe9949e1bab. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 841.535688] env[61728]: DEBUG oslo_concurrency.lockutils [req-5c635c12-afae-4523-8a08-31688891be10 req-fd92efab-cea7-49c8-808f-a991d181e019 service nova] Acquiring lock "refresh_cache-174607a1-9bc8-4e07-8993-7f0bb0f308e9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.535844] env[61728]: DEBUG oslo_concurrency.lockutils [req-5c635c12-afae-4523-8a08-31688891be10 req-fd92efab-cea7-49c8-808f-a991d181e019 service nova] Acquired lock "refresh_cache-174607a1-9bc8-4e07-8993-7f0bb0f308e9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.535922] env[61728]: DEBUG nova.network.neutron [req-5c635c12-afae-4523-8a08-31688891be10 req-fd92efab-cea7-49c8-808f-a991d181e019 service nova] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Refreshing network info cache for port 6969ab65-3bd4-4500-9908-9fe9949e1bab {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 841.557292] env[61728]: DEBUG oslo_concurrency.lockutils [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.112s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.559778] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.529s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.562345] env[61728]: INFO nova.compute.claims [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 841.577506] env[61728]: DEBUG oslo_vmware.api [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Task: {'id': task-464314, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.439274} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.578105] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.578359] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 841.578621] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 841.578852] env[61728]: INFO nova.compute.manager [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Took 1.44 seconds to destroy the instance on the hypervisor. [ 841.579211] env[61728]: DEBUG oslo.service.loopingcall [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.579469] env[61728]: DEBUG nova.compute.manager [-] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 841.579784] env[61728]: DEBUG nova.network.neutron [-] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 841.588155] env[61728]: INFO nova.scheduler.client.report [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Deleted allocations for instance a5c7e3ff-6829-4528-9279-e865a0eb8512 [ 841.613866] env[61728]: DEBUG oslo_vmware.api [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464311, 'name': PowerOnVM_Task, 'duration_secs': 0.972799} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.614210] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 841.614424] env[61728]: INFO nova.compute.manager [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Took 9.96 seconds to spawn the instance on the hypervisor. [ 841.614607] env[61728]: DEBUG nova.compute.manager [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.615424] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07881a76-5e67-45ca-a209-97a200818048 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.660072] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464315, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.660351] env[61728]: DEBUG oslo_vmware.api [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.663958] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 08ad3f4a-12a8-48d3-8829-f6744648a95e] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 842.004322] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c669b3-716d-0011-e53f-a4043be9094c, 'name': SearchDatastore_Task, 'duration_secs': 0.041244} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.005199] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-278e1463-88dc-4ade-b62e-c8ebec5668bf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.011374] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Waiting for the task: (returnval){ [ 842.011374] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c9bf18-48b9-e879-7c14-c4a7be8d3820" [ 842.011374] env[61728]: _type = "Task" [ 842.011374] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.023502] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c9bf18-48b9-e879-7c14-c4a7be8d3820, 'name': SearchDatastore_Task, 'duration_secs': 0.009813} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.024017] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.024231] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 174607a1-9bc8-4e07-8993-7f0bb0f308e9/174607a1-9bc8-4e07-8993-7f0bb0f308e9.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 842.024519] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2bac3a9f-3d9c-4d1c-9f9e-b52c1903bb76 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.034981] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Waiting for the task: (returnval){ [ 842.034981] env[61728]: value = "task-464317" [ 842.034981] env[61728]: _type = "Task" [ 842.034981] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.043273] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464317, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.105457] env[61728]: DEBUG oslo_concurrency.lockutils [None req-719f3ada-5dcb-44bc-893f-cc0551446b6a tempest-ServersTestJSON-179752495 tempest-ServersTestJSON-179752495-project-member] Lock "a5c7e3ff-6829-4528-9279-e865a0eb8512" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.827s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.141670] env[61728]: INFO nova.compute.manager [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Took 69.04 seconds to build instance. [ 842.167486] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464315, 'name': ReconfigVM_Task, 'duration_secs': 0.863009} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.167758] env[61728]: DEBUG oslo_vmware.api [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Task: {'id': task-464316, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.581479} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.168125] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 3cd416c6-9926-45ee-afe3-d3468e4e06c2] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 842.170587] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Reconfigured VM instance instance-00000046 to attach disk [datastore1] b444a6cf-3c98-4308-afc1-6e760d30082d/b444a6cf-3c98-4308-afc1-6e760d30082d.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.171228] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 842.171415] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 842.171731] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 842.171782] env[61728]: INFO nova.compute.manager [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Took 1.72 seconds to destroy the instance on the hypervisor. [ 842.172105] env[61728]: DEBUG oslo.service.loopingcall [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.172917] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a8d3265e-b3e7-4b01-8134-f6705df19733 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.174685] env[61728]: DEBUG nova.compute.manager [-] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 842.174725] env[61728]: DEBUG nova.network.neutron [-] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 842.182163] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 842.182163] env[61728]: value = "task-464318" [ 842.182163] env[61728]: _type = "Task" [ 842.182163] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.191653] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464318, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.488838] env[61728]: DEBUG nova.network.neutron [req-5c635c12-afae-4523-8a08-31688891be10 req-fd92efab-cea7-49c8-808f-a991d181e019 service nova] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Updated VIF entry in instance network info cache for port 6969ab65-3bd4-4500-9908-9fe9949e1bab. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 842.489039] env[61728]: DEBUG nova.network.neutron [req-5c635c12-afae-4523-8a08-31688891be10 req-fd92efab-cea7-49c8-808f-a991d181e019 service nova] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Updating instance_info_cache with network_info: [{"id": "6969ab65-3bd4-4500-9908-9fe9949e1bab", "address": "fa:16:3e:49:18:80", "network": {"id": "d24b80d4-9eda-45b0-b1bb-cd522a1260fc", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1872056846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9fe346266a245dbba86fb9c8cda0a94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3", "external-id": "nsx-vlan-transportzone-263", "segmentation_id": 263, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6969ab65-3b", "ovs_interfaceid": "6969ab65-3bd4-4500-9908-9fe9949e1bab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.543793] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464317, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.644128] env[61728]: DEBUG nova.network.neutron [-] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.645383] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a7d55d61-76f1-48be-934c-5482a2071f4a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "d44718bd-872b-401c-aa11-f10bea4a35d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.970s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.674307] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: d75ab8eb-f15c-4e0f-8376-6d91aa26ab58] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 842.694954] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464318, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.772078] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "121bef05-ef7b-47ee-b737-15f8b9d91158" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.772346] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "121bef05-ef7b-47ee-b737-15f8b9d91158" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.804018] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "40558dd0-69d3-42d6-87cf-8e4ae4ae9dae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.804355] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "40558dd0-69d3-42d6-87cf-8e4ae4ae9dae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.991898] env[61728]: DEBUG oslo_concurrency.lockutils [req-5c635c12-afae-4523-8a08-31688891be10 req-fd92efab-cea7-49c8-808f-a991d181e019 service nova] Releasing lock "refresh_cache-174607a1-9bc8-4e07-8993-7f0bb0f308e9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.998995] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb078ea-2bae-4273-b88c-afadf3598998 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.004509] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "interface-f28348d2-c062-497a-b374-521df51054ee-b32267f3-fc0e-4870-b60d-664d82e90ca7" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.004509] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-f28348d2-c062-497a-b374-521df51054ee-b32267f3-fc0e-4870-b60d-664d82e90ca7" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.004688] env[61728]: DEBUG nova.objects.instance [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'flavor' on Instance uuid f28348d2-c062-497a-b374-521df51054ee {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.009858] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb20e35-8572-46e4-8252-e860087b7a9b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.044203] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45564c7-ba18-4a2e-8708-25918b8ec1de {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.054547] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464317, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.059342] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5849d48-53d9-43b4-97da-c33c080a3272 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.076833] env[61728]: DEBUG nova.compute.provider_tree [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.085963] env[61728]: DEBUG nova.network.neutron [-] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.153697] env[61728]: INFO nova.compute.manager [-] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Took 1.57 seconds to deallocate network for instance. [ 843.177555] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 843.177746] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Cleaning up deleted instances with incomplete migration {{(pid=61728) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 843.200573] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464318, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.275538] env[61728]: DEBUG nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.306927] env[61728]: DEBUG nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.554824] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464317, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.322667} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.555046] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 174607a1-9bc8-4e07-8993-7f0bb0f308e9/174607a1-9bc8-4e07-8993-7f0bb0f308e9.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 843.555273] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 843.555532] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b41e3834-5d44-49cc-9705-7b7c3dd5c28c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.561661] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Waiting for the task: (returnval){ [ 843.561661] env[61728]: value = "task-464319" [ 843.561661] env[61728]: _type = "Task" [ 843.561661] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.571783] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464319, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.579831] env[61728]: DEBUG nova.scheduler.client.report [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 843.588123] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "f27f22aa-f05d-4981-a389-311731c51f93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.588355] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "f27f22aa-f05d-4981-a389-311731c51f93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.589664] env[61728]: INFO nova.compute.manager [-] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Took 1.41 seconds to deallocate network for instance. [ 843.661054] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.680194] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 843.698304] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464318, 'name': Rename_Task, 'duration_secs': 1.284952} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.698587] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 843.699263] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be665cf5-a5b4-48fc-ad14-bb00f4f8c3a8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.705600] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 843.705600] env[61728]: value = "task-464320" [ 843.705600] env[61728]: _type = "Task" [ 843.705600] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.714093] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464320, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.740285] env[61728]: DEBUG nova.objects.instance [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'pci_requests' on Instance uuid f28348d2-c062-497a-b374-521df51054ee {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.809582] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.831865] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.073735] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464319, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07202} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.074026] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 844.074901] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09679ea5-5216-4ecb-9da3-ac0c33656105 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.089978] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.530s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.092695] env[61728]: DEBUG nova.compute.manager [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 844.109583] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 174607a1-9bc8-4e07-8993-7f0bb0f308e9/174607a1-9bc8-4e07-8993-7f0bb0f308e9.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 844.110380] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.267s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.110663] env[61728]: DEBUG nova.objects.instance [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lazy-loading 'resources' on Instance uuid 1d9114cb-a0af-4874-962f-27237b3c89cc {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.114572] env[61728]: DEBUG nova.compute.manager [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 844.115192] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.115998] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-991c9291-7983-4b21-8f9c-d5366075bb72 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.137472] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Waiting for the task: (returnval){ [ 844.137472] env[61728]: value = "task-464321" [ 844.137472] env[61728]: _type = "Task" [ 844.137472] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.148389] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464321, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.219963] env[61728]: DEBUG oslo_vmware.api [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464320, 'name': PowerOnVM_Task, 'duration_secs': 0.485822} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.220487] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 844.220859] env[61728]: INFO nova.compute.manager [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Took 9.96 seconds to spawn the instance on the hypervisor. [ 844.221131] env[61728]: DEBUG nova.compute.manager [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 844.222310] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8ab9e7-46cb-4ffe-8a34-6f8d7cfc2ea3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.243115] env[61728]: DEBUG nova.objects.base [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 844.243370] env[61728]: DEBUG nova.network.neutron [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 844.355496] env[61728]: DEBUG nova.policy [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e51efcf23db4d5496edbc3b7104b4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32bb35e1dfce40e48be08bb568d3f2b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 844.614429] env[61728]: DEBUG nova.compute.utils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 844.615946] env[61728]: DEBUG nova.compute.manager [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 844.619508] env[61728]: DEBUG nova.network.neutron [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 844.653061] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464321, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.655639] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.714362] env[61728]: DEBUG nova.policy [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93f99e0c073647c08e8a0a082894c44d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73f71bf3a4534155af5c5a0255e38a4c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 844.742046] env[61728]: INFO nova.compute.manager [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Took 60.53 seconds to build instance. [ 844.842347] env[61728]: DEBUG nova.compute.manager [req-e272b8a3-26d9-4ec0-bbd9-85de33efae13 req-4e2b5baf-9b3b-40f7-be70-555939e07865 service nova] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Received event network-vif-deleted-cc02ac93-777e-4bcf-8af5-69f162374735 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 844.843463] env[61728]: DEBUG nova.compute.manager [req-e272b8a3-26d9-4ec0-bbd9-85de33efae13 req-4e2b5baf-9b3b-40f7-be70-555939e07865 service nova] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Received event network-vif-deleted-ebbb109a-f3c6-4a01-a162-7d8bca2c868c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 845.033978] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872645dc-0bcb-4591-b39d-c546a67605c4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.041265] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf664087-a2f0-4a45-8492-b17f418517bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.077713] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12309d0f-66db-41ce-a855-e0e4e22f56fe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.092021] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21569c5d-0518-49cc-9e46-211628bc519b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.102493] env[61728]: DEBUG nova.compute.provider_tree [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.120269] env[61728]: DEBUG nova.compute.manager [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 845.152811] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464321, 'name': ReconfigVM_Task, 'duration_secs': 0.813743} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.152811] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 174607a1-9bc8-4e07-8993-7f0bb0f308e9/174607a1-9bc8-4e07-8993-7f0bb0f308e9.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 845.153434] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-18ad4b88-5255-4237-a803-169e0f191b06 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.160759] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Waiting for the task: (returnval){ [ 845.160759] env[61728]: value = "task-464322" [ 845.160759] env[61728]: _type = "Task" [ 845.160759] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.172639] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464322, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.234969] env[61728]: DEBUG nova.network.neutron [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Successfully created port: 3bd62766-6bca-43f7-b2a6-fb3f63eee1c7 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 845.245613] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10e95f76-7d25-4875-8865-2cdd3f22d9cc tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "b444a6cf-3c98-4308-afc1-6e760d30082d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.046s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.605611] env[61728]: DEBUG nova.scheduler.client.report [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.673474] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464322, 'name': Rename_Task, 'duration_secs': 0.480312} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.673865] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 845.674306] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf3e49eb-ed37-4df8-97be-28370bfd4e76 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.681750] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Waiting for the task: (returnval){ [ 845.681750] env[61728]: value = "task-464323" [ 845.681750] env[61728]: _type = "Task" [ 845.681750] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.693685] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464323, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.110765] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.113908] env[61728]: DEBUG oslo_concurrency.lockutils [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.043s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.114357] env[61728]: DEBUG nova.objects.instance [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Lazy-loading 'resources' on Instance uuid e3116a87-ce12-4ec8-b6fa-70ec2127ebb0 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.134395] env[61728]: DEBUG nova.compute.manager [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 846.153235] env[61728]: INFO nova.scheduler.client.report [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Deleted allocations for instance 1d9114cb-a0af-4874-962f-27237b3c89cc [ 846.177611] env[61728]: DEBUG nova.virt.hardware [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 846.180030] env[61728]: DEBUG nova.virt.hardware [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 846.180030] env[61728]: DEBUG nova.virt.hardware [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 846.180030] env[61728]: DEBUG nova.virt.hardware [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 846.180030] env[61728]: DEBUG nova.virt.hardware [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 846.180030] env[61728]: DEBUG nova.virt.hardware [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 846.180030] env[61728]: DEBUG nova.virt.hardware [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 846.180030] env[61728]: DEBUG nova.virt.hardware [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 846.181702] env[61728]: DEBUG nova.virt.hardware [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 846.182058] env[61728]: DEBUG nova.virt.hardware [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 846.183381] env[61728]: DEBUG nova.virt.hardware [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 846.183478] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4bcdc5b-1eff-42eb-99ff-241c9e73f628 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.204930] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aee7b0d-075a-41b3-a18c-41b79506400e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.210087] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464323, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.293572] env[61728]: DEBUG nova.network.neutron [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Successfully updated port: b32267f3-fc0e-4870-b60d-664d82e90ca7 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 846.386990] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "5fa2e9f4-6d68-46c2-a549-95d216bab886" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.387266] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "5fa2e9f4-6d68-46c2-a549-95d216bab886" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.665482] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9c2df4e5-cd1c-4ab2-910a-4863d564139c tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "1d9114cb-a0af-4874-962f-27237b3c89cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.173s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.698945] env[61728]: DEBUG oslo_vmware.api [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464323, 'name': PowerOnVM_Task, 'duration_secs': 0.536314} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.701696] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 846.701941] env[61728]: INFO nova.compute.manager [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Took 9.89 seconds to spawn the instance on the hypervisor. [ 846.702143] env[61728]: DEBUG nova.compute.manager [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.704020] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb1673f-12c8-4bbf-b308-2cb1cd069c20 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.799267] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.799615] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.799880] env[61728]: DEBUG nova.network.neutron [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 846.890294] env[61728]: DEBUG nova.compute.manager [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 847.069623] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0416c6de-0ad3-4c08-9093-f6417216ad99 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.077726] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e74696cd-3d63-4e20-abef-f423057c7965 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.112159] env[61728]: DEBUG nova.network.neutron [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Successfully updated port: 3bd62766-6bca-43f7-b2a6-fb3f63eee1c7 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 847.116564] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b954a5a-dc93-4941-847f-711c255957d0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.121859] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1f5fc5-dd50-484e-af59-6cd6a89378f7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.136982] env[61728]: DEBUG nova.compute.provider_tree [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.224657] env[61728]: INFO nova.compute.manager [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Took 52.49 seconds to build instance. [ 847.350178] env[61728]: WARNING nova.network.neutron [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] 20ba9b61-3be9-4ad2-a1d0-88c810873bcb already exists in list: networks containing: ['20ba9b61-3be9-4ad2-a1d0-88c810873bcb']. ignoring it [ 847.350178] env[61728]: WARNING nova.network.neutron [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] 20ba9b61-3be9-4ad2-a1d0-88c810873bcb already exists in list: networks containing: ['20ba9b61-3be9-4ad2-a1d0-88c810873bcb']. ignoring it [ 847.418833] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.617261] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Acquiring lock "refresh_cache-a9745dc8-6400-424c-bc10-1181ee128648" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.617421] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Acquired lock "refresh_cache-a9745dc8-6400-424c-bc10-1181ee128648" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.617575] env[61728]: DEBUG nova.network.neutron [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 847.642965] env[61728]: DEBUG nova.scheduler.client.report [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.728466] env[61728]: DEBUG oslo_concurrency.lockutils [None req-42c40903-15f7-4793-8762-3b95cde8dc6b tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Lock "174607a1-9bc8-4e07-8993-7f0bb0f308e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.004s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.063362] env[61728]: DEBUG oslo_concurrency.lockutils [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "f2692e22-6c14-4bb3-a080-607f0731105d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.063642] env[61728]: DEBUG oslo_concurrency.lockutils [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "f2692e22-6c14-4bb3-a080-607f0731105d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.063877] env[61728]: DEBUG oslo_concurrency.lockutils [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "f2692e22-6c14-4bb3-a080-607f0731105d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.064104] env[61728]: DEBUG oslo_concurrency.lockutils [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "f2692e22-6c14-4bb3-a080-607f0731105d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.064362] env[61728]: DEBUG oslo_concurrency.lockutils [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "f2692e22-6c14-4bb3-a080-607f0731105d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.066688] env[61728]: INFO nova.compute.manager [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Terminating instance [ 848.068516] env[61728]: DEBUG nova.compute.manager [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 848.068726] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 848.071783] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ad4d60-436d-49c3-bc77-8819937e90df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.083739] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 848.083739] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06f81ac1-e38c-4c8e-9e85-4ed1440193dc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.095691] env[61728]: DEBUG oslo_vmware.api [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 848.095691] env[61728]: value = "task-464324" [ 848.095691] env[61728]: _type = "Task" [ 848.095691] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.105027] env[61728]: DEBUG oslo_vmware.api [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464324, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.149169] env[61728]: DEBUG oslo_concurrency.lockutils [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.035s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.152853] env[61728]: DEBUG oslo_concurrency.lockutils [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.037s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.153090] env[61728]: DEBUG nova.objects.instance [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61728) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 848.177670] env[61728]: DEBUG nova.network.neutron [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 848.189695] env[61728]: INFO nova.scheduler.client.report [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Deleted allocations for instance e3116a87-ce12-4ec8-b6fa-70ec2127ebb0 [ 848.349781] env[61728]: DEBUG nova.network.neutron [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Updating instance_info_cache with network_info: [{"id": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "address": "fa:16:3e:44:e0:e6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e719ade-f3", "ovs_interfaceid": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "02293cc2-3b2d-4b93-b27d-0e61b2780e89", "address": "fa:16:3e:39:43:bf", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02293cc2-3b", "ovs_interfaceid": "02293cc2-3b2d-4b93-b27d-0e61b2780e89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b32267f3-fc0e-4870-b60d-664d82e90ca7", "address": "fa:16:3e:99:4f:52", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb32267f3-fc", "ovs_interfaceid": "b32267f3-fc0e-4870-b60d-664d82e90ca7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.373254] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "d44718bd-872b-401c-aa11-f10bea4a35d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.373254] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "d44718bd-872b-401c-aa11-f10bea4a35d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.373532] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "d44718bd-872b-401c-aa11-f10bea4a35d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.373938] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "d44718bd-872b-401c-aa11-f10bea4a35d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.374168] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "d44718bd-872b-401c-aa11-f10bea4a35d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.376250] env[61728]: INFO nova.compute.manager [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Terminating instance [ 848.378118] env[61728]: DEBUG nova.compute.manager [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 848.378560] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 848.379550] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4812307f-eab0-4358-9c3a-2062d2de5c88 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.388805] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 848.389070] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f83108e-cc66-42ac-86d0-8e13cdcba9df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.401984] env[61728]: DEBUG oslo_vmware.api [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 848.401984] env[61728]: value = "task-464325" [ 848.401984] env[61728]: _type = "Task" [ 848.401984] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.408121] env[61728]: DEBUG nova.compute.manager [req-fe8b1384-9444-471f-86de-adeabd66ebc2 req-288680a5-8f4d-4a10-ac9e-c2b952ca8189 service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Received event network-vif-plugged-b32267f3-fc0e-4870-b60d-664d82e90ca7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 848.408351] env[61728]: DEBUG oslo_concurrency.lockutils [req-fe8b1384-9444-471f-86de-adeabd66ebc2 req-288680a5-8f4d-4a10-ac9e-c2b952ca8189 service nova] Acquiring lock "f28348d2-c062-497a-b374-521df51054ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.408658] env[61728]: DEBUG oslo_concurrency.lockutils [req-fe8b1384-9444-471f-86de-adeabd66ebc2 req-288680a5-8f4d-4a10-ac9e-c2b952ca8189 service nova] Lock "f28348d2-c062-497a-b374-521df51054ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.408728] env[61728]: DEBUG oslo_concurrency.lockutils [req-fe8b1384-9444-471f-86de-adeabd66ebc2 req-288680a5-8f4d-4a10-ac9e-c2b952ca8189 service nova] Lock "f28348d2-c062-497a-b374-521df51054ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.408919] env[61728]: DEBUG nova.compute.manager [req-fe8b1384-9444-471f-86de-adeabd66ebc2 req-288680a5-8f4d-4a10-ac9e-c2b952ca8189 service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] No waiting events found dispatching network-vif-plugged-b32267f3-fc0e-4870-b60d-664d82e90ca7 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 848.409103] env[61728]: WARNING nova.compute.manager [req-fe8b1384-9444-471f-86de-adeabd66ebc2 req-288680a5-8f4d-4a10-ac9e-c2b952ca8189 service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Received unexpected event network-vif-plugged-b32267f3-fc0e-4870-b60d-664d82e90ca7 for instance with vm_state active and task_state None. [ 848.409255] env[61728]: DEBUG nova.compute.manager [req-fe8b1384-9444-471f-86de-adeabd66ebc2 req-288680a5-8f4d-4a10-ac9e-c2b952ca8189 service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Received event network-changed-b32267f3-fc0e-4870-b60d-664d82e90ca7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 848.409418] env[61728]: DEBUG nova.compute.manager [req-fe8b1384-9444-471f-86de-adeabd66ebc2 req-288680a5-8f4d-4a10-ac9e-c2b952ca8189 service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Refreshing instance network info cache due to event network-changed-b32267f3-fc0e-4870-b60d-664d82e90ca7. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 848.409589] env[61728]: DEBUG oslo_concurrency.lockutils [req-fe8b1384-9444-471f-86de-adeabd66ebc2 req-288680a5-8f4d-4a10-ac9e-c2b952ca8189 service nova] Acquiring lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.421140] env[61728]: DEBUG oslo_vmware.api [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464325, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.467339] env[61728]: DEBUG nova.network.neutron [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Updating instance_info_cache with network_info: [{"id": "3bd62766-6bca-43f7-b2a6-fb3f63eee1c7", "address": "fa:16:3e:36:9b:25", "network": {"id": "88b88845-2c7f-4bbd-b77e-b09e1580599d", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1136388705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73f71bf3a4534155af5c5a0255e38a4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bd62766-6b", "ovs_interfaceid": "3bd62766-6bca-43f7-b2a6-fb3f63eee1c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.610624] env[61728]: DEBUG oslo_vmware.api [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464324, 'name': PowerOffVM_Task, 'duration_secs': 0.418115} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.611619] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 848.611619] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 848.611619] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03e8e54b-2829-4d6c-b7ad-77b3df2a04fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.686016] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 848.686913] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 848.687356] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleting the datastore file [datastore1] f2692e22-6c14-4bb3-a080-607f0731105d {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 848.687621] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-76fa1145-1f90-4e04-a39a-cd182834729b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.701527] env[61728]: DEBUG oslo_vmware.api [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 848.701527] env[61728]: value = "task-464327" [ 848.701527] env[61728]: _type = "Task" [ 848.701527] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.702112] env[61728]: DEBUG oslo_concurrency.lockutils [None req-458ed6e2-7e51-444b-a6ce-f8bbef7e1f0b tempest-AttachInterfacesUnderV243Test-1372391865 tempest-AttachInterfacesUnderV243Test-1372391865-project-member] Lock "e3116a87-ce12-4ec8-b6fa-70ec2127ebb0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.107s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.715427] env[61728]: DEBUG oslo_vmware.api [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464327, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.854358] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.856050] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.856050] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.856326] env[61728]: DEBUG oslo_concurrency.lockutils [req-fe8b1384-9444-471f-86de-adeabd66ebc2 req-288680a5-8f4d-4a10-ac9e-c2b952ca8189 service nova] Acquired lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.856430] env[61728]: DEBUG nova.network.neutron [req-fe8b1384-9444-471f-86de-adeabd66ebc2 req-288680a5-8f4d-4a10-ac9e-c2b952ca8189 service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Refreshing network info cache for port b32267f3-fc0e-4870-b60d-664d82e90ca7 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 848.858730] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89141f2d-b17d-4193-9e94-1e668af83267 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.881208] env[61728]: DEBUG nova.virt.hardware [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 848.881691] env[61728]: DEBUG nova.virt.hardware [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 848.881946] env[61728]: DEBUG nova.virt.hardware [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 848.882171] env[61728]: DEBUG nova.virt.hardware [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 848.882330] env[61728]: DEBUG nova.virt.hardware [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 848.882523] env[61728]: DEBUG nova.virt.hardware [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 848.882786] env[61728]: DEBUG nova.virt.hardware [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 848.882969] env[61728]: DEBUG nova.virt.hardware [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 848.883195] env[61728]: DEBUG nova.virt.hardware [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 848.883377] env[61728]: DEBUG nova.virt.hardware [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 848.883571] env[61728]: DEBUG nova.virt.hardware [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 848.892829] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Reconfiguring VM to attach interface {{(pid=61728) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 848.893636] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32504a42-4a08-47f5-b4d6-ba4aee755484 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.921924] env[61728]: DEBUG oslo_vmware.api [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464325, 'name': PowerOffVM_Task, 'duration_secs': 0.210337} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.923563] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 848.923791] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 848.924202] env[61728]: DEBUG oslo_vmware.api [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 848.924202] env[61728]: value = "task-464328" [ 848.924202] env[61728]: _type = "Task" [ 848.924202] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.924432] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2805524-9f1a-47d4-8d04-642dbaaa0c69 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.941062] env[61728]: DEBUG oslo_vmware.api [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464328, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.970030] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Releasing lock "refresh_cache-a9745dc8-6400-424c-bc10-1181ee128648" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.970261] env[61728]: DEBUG nova.compute.manager [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Instance network_info: |[{"id": "3bd62766-6bca-43f7-b2a6-fb3f63eee1c7", "address": "fa:16:3e:36:9b:25", "network": {"id": "88b88845-2c7f-4bbd-b77e-b09e1580599d", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1136388705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73f71bf3a4534155af5c5a0255e38a4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bd62766-6b", "ovs_interfaceid": "3bd62766-6bca-43f7-b2a6-fb3f63eee1c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 848.972047] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:9b:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c3e2368-4a35-4aa5-9135-23daedbbf9ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3bd62766-6bca-43f7-b2a6-fb3f63eee1c7', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 848.980219] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Creating folder: Project (73f71bf3a4534155af5c5a0255e38a4c). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 848.981021] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-64d210d8-43f8-40e7-b799-3b79a3607013 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.995639] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Created folder: Project (73f71bf3a4534155af5c5a0255e38a4c) in parent group-v121913. [ 848.996438] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Creating folder: Instances. Parent ref: group-v122123. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 848.996438] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-edffb384-9ae2-4318-8d75-f1f5ff01797f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.001197] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 849.001197] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 849.001409] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleting the datastore file [datastore1] d44718bd-872b-401c-aa11-f10bea4a35d8 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 849.002247] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a0738e98-9a66-43ce-baf2-d1ce7fb54395 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.015856] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Created folder: Instances in parent group-v122123. [ 849.016166] env[61728]: DEBUG oslo.service.loopingcall [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.016384] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 849.016602] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d1e20a72-cce3-4ceb-9ea6-dd08feb55eb4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.033256] env[61728]: DEBUG oslo_vmware.api [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 849.033256] env[61728]: value = "task-464332" [ 849.033256] env[61728]: _type = "Task" [ 849.033256] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.040279] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.040279] env[61728]: value = "task-464333" [ 849.040279] env[61728]: _type = "Task" [ 849.040279] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.044813] env[61728]: DEBUG oslo_vmware.api [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464332, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.051541] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464333, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.172476] env[61728]: DEBUG oslo_concurrency.lockutils [None req-648f480d-87f3-45a6-ac08-89d4ead716a6 tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.173713] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.834s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.175229] env[61728]: INFO nova.compute.claims [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 849.216706] env[61728]: DEBUG oslo_vmware.api [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464327, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179087} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.216706] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 849.216706] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 849.217050] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 849.217124] env[61728]: INFO nova.compute.manager [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Took 1.15 seconds to destroy the instance on the hypervisor. [ 849.217334] env[61728]: DEBUG oslo.service.loopingcall [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.217573] env[61728]: DEBUG nova.compute.manager [-] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 849.217666] env[61728]: DEBUG nova.network.neutron [-] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 849.439107] env[61728]: DEBUG oslo_vmware.api [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.551474] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "3415687e-9b8d-42f8-9b18-93951b8ec7da" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.551749] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "3415687e-9b8d-42f8-9b18-93951b8ec7da" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.560703] env[61728]: DEBUG oslo_vmware.api [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464332, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254533} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.560703] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 849.560703] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 849.560703] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 849.560703] env[61728]: INFO nova.compute.manager [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Took 1.18 seconds to destroy the instance on the hypervisor. [ 849.560703] env[61728]: DEBUG oslo.service.loopingcall [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.563281] env[61728]: DEBUG nova.compute.manager [-] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 849.563407] env[61728]: DEBUG nova.network.neutron [-] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 849.573723] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464333, 'name': CreateVM_Task, 'duration_secs': 0.476442} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.575766] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 849.576454] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.576635] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.576951] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 849.577831] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f58898bb-4e40-4874-a386-8afa9d662c33 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.585133] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Waiting for the task: (returnval){ [ 849.585133] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aa5980-de9b-4223-f980-646fe8406b08" [ 849.585133] env[61728]: _type = "Task" [ 849.585133] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.595273] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aa5980-de9b-4223-f980-646fe8406b08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.810297] env[61728]: DEBUG nova.network.neutron [req-fe8b1384-9444-471f-86de-adeabd66ebc2 req-288680a5-8f4d-4a10-ac9e-c2b952ca8189 service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Updated VIF entry in instance network info cache for port b32267f3-fc0e-4870-b60d-664d82e90ca7. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 849.810297] env[61728]: DEBUG nova.network.neutron [req-fe8b1384-9444-471f-86de-adeabd66ebc2 req-288680a5-8f4d-4a10-ac9e-c2b952ca8189 service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Updating instance_info_cache with network_info: [{"id": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "address": "fa:16:3e:44:e0:e6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e719ade-f3", "ovs_interfaceid": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "02293cc2-3b2d-4b93-b27d-0e61b2780e89", "address": "fa:16:3e:39:43:bf", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02293cc2-3b", "ovs_interfaceid": "02293cc2-3b2d-4b93-b27d-0e61b2780e89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b32267f3-fc0e-4870-b60d-664d82e90ca7", "address": "fa:16:3e:99:4f:52", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb32267f3-fc", "ovs_interfaceid": "b32267f3-fc0e-4870-b60d-664d82e90ca7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.938150] env[61728]: DEBUG oslo_vmware.api [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464328, 'name': ReconfigVM_Task, 'duration_secs': 0.868666} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.938668] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.938982] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Reconfigured VM to attach interface {{(pid=61728) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 850.061711] env[61728]: DEBUG nova.compute.manager [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 850.102153] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aa5980-de9b-4223-f980-646fe8406b08, 'name': SearchDatastore_Task, 'duration_secs': 0.017316} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.102714] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.102956] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.103212] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.103368] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.103550] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.103857] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a61b13fc-ddc4-4688-ab8c-0be5ff7051dc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.115543] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.115741] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 850.116702] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab8e8150-5428-44f9-aef4-d99e717f1e28 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.122899] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Waiting for the task: (returnval){ [ 850.122899] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5286f922-6d77-34d2-e321-596415988111" [ 850.122899] env[61728]: _type = "Task" [ 850.122899] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.131322] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5286f922-6d77-34d2-e321-596415988111, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.285872] env[61728]: DEBUG nova.network.neutron [-] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.314439] env[61728]: DEBUG oslo_concurrency.lockutils [req-fe8b1384-9444-471f-86de-adeabd66ebc2 req-288680a5-8f4d-4a10-ac9e-c2b952ca8189 service nova] Releasing lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.446102] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cffbde3c-e63a-4925-b034-def3576805d1 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-f28348d2-c062-497a-b374-521df51054ee-b32267f3-fc0e-4870-b60d-664d82e90ca7" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.442s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.547117] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78e3994-9ed8-4832-985e-bb4a82687ef1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.555661] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1d57df-3ebc-405b-a2b4-7af36bafe3ee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.592654] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba2ae2c-1124-478a-9feb-cb9ffd03a97f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.601203] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57cb3be3-6c0b-40c3-9efc-4eac1dcb4673 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.606072] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.606367] env[61728]: DEBUG nova.network.neutron [-] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.619691] env[61728]: DEBUG nova.compute.provider_tree [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.634981] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5286f922-6d77-34d2-e321-596415988111, 'name': SearchDatastore_Task, 'duration_secs': 0.030212} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.635595] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9447fe56-8f6d-4c22-84ca-a33b8fa153a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.641670] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Waiting for the task: (returnval){ [ 850.641670] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5223b582-3203-0fc3-0ce8-0d133c3ee223" [ 850.641670] env[61728]: _type = "Task" [ 850.641670] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.650486] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5223b582-3203-0fc3-0ce8-0d133c3ee223, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.789283] env[61728]: INFO nova.compute.manager [-] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Took 1.57 seconds to deallocate network for instance. [ 850.998088] env[61728]: DEBUG oslo_concurrency.lockutils [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "b444a6cf-3c98-4308-afc1-6e760d30082d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.998088] env[61728]: DEBUG oslo_concurrency.lockutils [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "b444a6cf-3c98-4308-afc1-6e760d30082d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.998284] env[61728]: DEBUG oslo_concurrency.lockutils [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "b444a6cf-3c98-4308-afc1-6e760d30082d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.998494] env[61728]: DEBUG oslo_concurrency.lockutils [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "b444a6cf-3c98-4308-afc1-6e760d30082d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.000200] env[61728]: DEBUG oslo_concurrency.lockutils [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "b444a6cf-3c98-4308-afc1-6e760d30082d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.002s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.003150] env[61728]: INFO nova.compute.manager [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Terminating instance [ 851.005277] env[61728]: DEBUG nova.compute.manager [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 851.005410] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 851.006314] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-111b50cc-d8a2-464f-a3b7-c5610b083d30 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.015190] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 851.015478] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ab523b0b-a8a1-4a9d-b21a-952690c4f39c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.023894] env[61728]: DEBUG oslo_vmware.api [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 851.023894] env[61728]: value = "task-464334" [ 851.023894] env[61728]: _type = "Task" [ 851.023894] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.036435] env[61728]: DEBUG oslo_vmware.api [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464334, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.110753] env[61728]: INFO nova.compute.manager [-] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Took 1.55 seconds to deallocate network for instance. [ 851.112106] env[61728]: DEBUG nova.compute.manager [req-50b827f6-1b98-4daf-9a0f-8c99e066e817 req-de224d38-09b1-4548-9aa9-6a12eea9cfad service nova] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Received event network-changed-6969ab65-3bd4-4500-9908-9fe9949e1bab {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 851.112305] env[61728]: DEBUG nova.compute.manager [req-50b827f6-1b98-4daf-9a0f-8c99e066e817 req-de224d38-09b1-4548-9aa9-6a12eea9cfad service nova] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Refreshing instance network info cache due to event network-changed-6969ab65-3bd4-4500-9908-9fe9949e1bab. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 851.112522] env[61728]: DEBUG oslo_concurrency.lockutils [req-50b827f6-1b98-4daf-9a0f-8c99e066e817 req-de224d38-09b1-4548-9aa9-6a12eea9cfad service nova] Acquiring lock "refresh_cache-174607a1-9bc8-4e07-8993-7f0bb0f308e9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.112698] env[61728]: DEBUG oslo_concurrency.lockutils [req-50b827f6-1b98-4daf-9a0f-8c99e066e817 req-de224d38-09b1-4548-9aa9-6a12eea9cfad service nova] Acquired lock "refresh_cache-174607a1-9bc8-4e07-8993-7f0bb0f308e9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.112882] env[61728]: DEBUG nova.network.neutron [req-50b827f6-1b98-4daf-9a0f-8c99e066e817 req-de224d38-09b1-4548-9aa9-6a12eea9cfad service nova] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Refreshing network info cache for port 6969ab65-3bd4-4500-9908-9fe9949e1bab {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 851.125202] env[61728]: DEBUG nova.scheduler.client.report [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.157531] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5223b582-3203-0fc3-0ce8-0d133c3ee223, 'name': SearchDatastore_Task, 'duration_secs': 0.01463} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.157886] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.158195] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a9745dc8-6400-424c-bc10-1181ee128648/a9745dc8-6400-424c-bc10-1181ee128648.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 851.158527] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5f3cc2aa-fb9e-437b-8d2b-f612f75f8d5a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.168289] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Waiting for the task: (returnval){ [ 851.168289] env[61728]: value = "task-464335" [ 851.168289] env[61728]: _type = "Task" [ 851.168289] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.180180] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464335, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.296392] env[61728]: DEBUG oslo_concurrency.lockutils [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.415259] env[61728]: DEBUG nova.compute.manager [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Received event network-vif-plugged-3bd62766-6bca-43f7-b2a6-fb3f63eee1c7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 851.415607] env[61728]: DEBUG oslo_concurrency.lockutils [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] Acquiring lock "a9745dc8-6400-424c-bc10-1181ee128648-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.415762] env[61728]: DEBUG oslo_concurrency.lockutils [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] Lock "a9745dc8-6400-424c-bc10-1181ee128648-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.415906] env[61728]: DEBUG oslo_concurrency.lockutils [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] Lock "a9745dc8-6400-424c-bc10-1181ee128648-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.416089] env[61728]: DEBUG nova.compute.manager [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] [instance: a9745dc8-6400-424c-bc10-1181ee128648] No waiting events found dispatching network-vif-plugged-3bd62766-6bca-43f7-b2a6-fb3f63eee1c7 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 851.416342] env[61728]: WARNING nova.compute.manager [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Received unexpected event network-vif-plugged-3bd62766-6bca-43f7-b2a6-fb3f63eee1c7 for instance with vm_state building and task_state spawning. [ 851.416418] env[61728]: DEBUG nova.compute.manager [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Received event network-changed-3bd62766-6bca-43f7-b2a6-fb3f63eee1c7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 851.419226] env[61728]: DEBUG nova.compute.manager [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Refreshing instance network info cache due to event network-changed-3bd62766-6bca-43f7-b2a6-fb3f63eee1c7. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 851.419226] env[61728]: DEBUG oslo_concurrency.lockutils [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] Acquiring lock "refresh_cache-a9745dc8-6400-424c-bc10-1181ee128648" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.419226] env[61728]: DEBUG oslo_concurrency.lockutils [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] Acquired lock "refresh_cache-a9745dc8-6400-424c-bc10-1181ee128648" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.419226] env[61728]: DEBUG nova.network.neutron [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Refreshing network info cache for port 3bd62766-6bca-43f7-b2a6-fb3f63eee1c7 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 851.541851] env[61728]: DEBUG oslo_vmware.api [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464334, 'name': PowerOffVM_Task, 'duration_secs': 0.3139} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.542209] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 851.542388] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 851.542656] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4deada0-7c3f-4e6d-91ea-f7efddcf110c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.584155] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Acquiring lock "fef6cb9e-d8c7-45d3-a7f6-e361ff704006" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.586030] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Lock "fef6cb9e-d8c7-45d3-a7f6-e361ff704006" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.622807] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.630193] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.631827] env[61728]: DEBUG nova.compute.manager [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 851.637117] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.890s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.637286] env[61728]: DEBUG nova.objects.instance [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'resources' on Instance uuid 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.654216] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 851.655196] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 851.655465] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Deleting the datastore file [datastore1] b444a6cf-3c98-4308-afc1-6e760d30082d {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 851.656854] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cd86e2d0-6837-4277-9a78-10206434280c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.675743] env[61728]: DEBUG oslo_vmware.api [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 851.675743] env[61728]: value = "task-464337" [ 851.675743] env[61728]: _type = "Task" [ 851.675743] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.687408] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464335, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.693650] env[61728]: DEBUG oslo_vmware.api [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464337, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.694037] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.694285] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.694521] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.694728] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.694925] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.697274] env[61728]: INFO nova.compute.manager [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Terminating instance [ 851.699257] env[61728]: DEBUG nova.compute.manager [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 851.699493] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 851.700482] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8ed7b6-84d8-4321-b04d-1e502b7d24fc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.709230] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 851.709491] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec608fd2-f1b9-402d-bbcc-c62c9a57d6fa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.717835] env[61728]: DEBUG oslo_vmware.api [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 851.717835] env[61728]: value = "task-464338" [ 851.717835] env[61728]: _type = "Task" [ 851.717835] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.728081] env[61728]: DEBUG oslo_vmware.api [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464338, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.088676] env[61728]: DEBUG nova.compute.manager [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 852.145703] env[61728]: DEBUG nova.compute.utils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 852.148743] env[61728]: DEBUG nova.compute.manager [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 852.149169] env[61728]: DEBUG nova.network.neutron [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 852.175221] env[61728]: DEBUG nova.network.neutron [req-50b827f6-1b98-4daf-9a0f-8c99e066e817 req-de224d38-09b1-4548-9aa9-6a12eea9cfad service nova] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Updated VIF entry in instance network info cache for port 6969ab65-3bd4-4500-9908-9fe9949e1bab. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 852.175221] env[61728]: DEBUG nova.network.neutron [req-50b827f6-1b98-4daf-9a0f-8c99e066e817 req-de224d38-09b1-4548-9aa9-6a12eea9cfad service nova] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Updating instance_info_cache with network_info: [{"id": "6969ab65-3bd4-4500-9908-9fe9949e1bab", "address": "fa:16:3e:49:18:80", "network": {"id": "d24b80d4-9eda-45b0-b1bb-cd522a1260fc", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1872056846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9fe346266a245dbba86fb9c8cda0a94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3", "external-id": "nsx-vlan-transportzone-263", "segmentation_id": 263, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6969ab65-3b", "ovs_interfaceid": "6969ab65-3bd4-4500-9908-9fe9949e1bab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.189613] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464335, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523132} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.191633] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a9745dc8-6400-424c-bc10-1181ee128648/a9745dc8-6400-424c-bc10-1181ee128648.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 852.192139] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.195226] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-673c15a8-0b7f-40c1-a6c2-f85d78ef9a5d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.207162] env[61728]: DEBUG oslo_vmware.api [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464337, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126665} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.207162] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 852.207162] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 852.207665] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 852.208010] env[61728]: INFO nova.compute.manager [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Took 1.20 seconds to destroy the instance on the hypervisor. [ 852.208416] env[61728]: DEBUG oslo.service.loopingcall [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.209227] env[61728]: DEBUG nova.compute.manager [-] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 852.209439] env[61728]: DEBUG nova.network.neutron [-] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 852.212407] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Waiting for the task: (returnval){ [ 852.212407] env[61728]: value = "task-464339" [ 852.212407] env[61728]: _type = "Task" [ 852.212407] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.222114] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464339, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.231190] env[61728]: DEBUG oslo_vmware.api [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464338, 'name': PowerOffVM_Task, 'duration_secs': 0.227452} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.233896] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 852.234132] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 852.235019] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6e19a009-aaf2-4157-b545-5ae675aa2226 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.251426] env[61728]: DEBUG nova.policy [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '091abd20ebfe450b9283f5c7df1c7993', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c78292171414530a1caba3dfd0f620f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 852.279715] env[61728]: DEBUG nova.network.neutron [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Updated VIF entry in instance network info cache for port 3bd62766-6bca-43f7-b2a6-fb3f63eee1c7. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 852.280094] env[61728]: DEBUG nova.network.neutron [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Updating instance_info_cache with network_info: [{"id": "3bd62766-6bca-43f7-b2a6-fb3f63eee1c7", "address": "fa:16:3e:36:9b:25", "network": {"id": "88b88845-2c7f-4bbd-b77e-b09e1580599d", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1136388705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73f71bf3a4534155af5c5a0255e38a4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bd62766-6b", "ovs_interfaceid": "3bd62766-6bca-43f7-b2a6-fb3f63eee1c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.311309] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 852.311703] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 852.311846] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Deleting the datastore file [datastore1] 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 852.315236] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf18f742-81f6-4233-9b01-daaa06c57237 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.323509] env[61728]: DEBUG oslo_vmware.api [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for the task: (returnval){ [ 852.323509] env[61728]: value = "task-464341" [ 852.323509] env[61728]: _type = "Task" [ 852.323509] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.332855] env[61728]: DEBUG oslo_vmware.api [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464341, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.540146] env[61728]: DEBUG nova.network.neutron [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Successfully created port: 12471ef1-0e45-46ee-9b2a-b5008289a370 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.569610] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733e2d69-e3f7-4a20-aae0-1fb2255327b9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.579948] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033ae419-791b-4bf9-9fbd-4e99fcff43c7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.621714] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d198f1-375a-4b3a-b14a-3192acbd51e4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.630697] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0466b1b8-75f0-47bf-892b-6b1a0413980d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.637681] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.648589] env[61728]: DEBUG nova.compute.provider_tree [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.652609] env[61728]: DEBUG nova.compute.manager [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 852.679224] env[61728]: DEBUG oslo_concurrency.lockutils [req-50b827f6-1b98-4daf-9a0f-8c99e066e817 req-de224d38-09b1-4548-9aa9-6a12eea9cfad service nova] Releasing lock "refresh_cache-174607a1-9bc8-4e07-8993-7f0bb0f308e9" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.724014] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464339, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069438} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.724423] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 852.725251] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c757e6b-3b3e-48b2-8834-ba5a27b43845 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.755022] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] a9745dc8-6400-424c-bc10-1181ee128648/a9745dc8-6400-424c-bc10-1181ee128648.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.755351] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8979b644-b626-48e1-8fdc-2636695c8494 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.777709] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Waiting for the task: (returnval){ [ 852.777709] env[61728]: value = "task-464342" [ 852.777709] env[61728]: _type = "Task" [ 852.777709] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.786713] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464342, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.788409] env[61728]: DEBUG oslo_concurrency.lockutils [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] Releasing lock "refresh_cache-a9745dc8-6400-424c-bc10-1181ee128648" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.788651] env[61728]: DEBUG nova.compute.manager [req-ff289664-6a7c-45e9-9b21-eb11b3b05879 req-647e747d-85d3-452c-89b0-80ba500ef5b3 service nova] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Received event network-vif-deleted-cc2e7ae1-f4d4-408e-bd5a-757fe8ae77bd {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 852.833718] env[61728]: DEBUG oslo_vmware.api [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Task: {'id': task-464341, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139596} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.834016] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 852.834318] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 852.834514] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 852.835322] env[61728]: INFO nova.compute.manager [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Took 1.14 seconds to destroy the instance on the hypervisor. [ 852.835609] env[61728]: DEBUG oslo.service.loopingcall [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.836179] env[61728]: DEBUG nova.compute.manager [-] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 852.836285] env[61728]: DEBUG nova.network.neutron [-] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 853.125613] env[61728]: DEBUG nova.network.neutron [-] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.149759] env[61728]: DEBUG nova.scheduler.client.report [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.290172] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464342, 'name': ReconfigVM_Task, 'duration_secs': 0.282762} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.290633] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Reconfigured VM instance instance-00000048 to attach disk [datastore1] a9745dc8-6400-424c-bc10-1181ee128648/a9745dc8-6400-424c-bc10-1181ee128648.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 853.292690] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-baeeb32e-e707-4cb3-aeea-ce0342134a1f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.303043] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Waiting for the task: (returnval){ [ 853.303043] env[61728]: value = "task-464343" [ 853.303043] env[61728]: _type = "Task" [ 853.303043] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.309409] env[61728]: DEBUG nova.compute.manager [req-8ac38a3c-e304-48cf-bd68-f96d564adcb5 req-2809ecb7-8645-482f-94a8-2784bb5cb4ec service nova] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Received event network-vif-deleted-1a5d5fc4-4ae6-467a-8bff-0c0f410c9252 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 853.309869] env[61728]: DEBUG nova.compute.manager [req-8ac38a3c-e304-48cf-bd68-f96d564adcb5 req-2809ecb7-8645-482f-94a8-2784bb5cb4ec service nova] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Received event network-vif-deleted-8904a7fb-a87a-4f80-958e-87748f00b59e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 853.317452] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464343, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.378246] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "interface-f28348d2-c062-497a-b374-521df51054ee-02293cc2-3b2d-4b93-b27d-0e61b2780e89" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.378476] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-f28348d2-c062-497a-b374-521df51054ee-02293cc2-3b2d-4b93-b27d-0e61b2780e89" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.628715] env[61728]: INFO nova.compute.manager [-] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Took 1.42 seconds to deallocate network for instance. [ 853.663474] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.026s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.666896] env[61728]: DEBUG oslo_concurrency.lockutils [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.487s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.667202] env[61728]: DEBUG nova.objects.instance [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Lazy-loading 'resources' on Instance uuid 6d55bb01-7855-4d86-ac57-ecae15ea590f {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 853.669092] env[61728]: DEBUG nova.compute.manager [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 853.686397] env[61728]: INFO nova.scheduler.client.report [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleted allocations for instance 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2 [ 853.699981] env[61728]: DEBUG nova.network.neutron [-] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.705854] env[61728]: DEBUG nova.virt.hardware [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='5b847cca62791d20b2aba41df2c080b1',container_format='bare',created_at=2024-08-29T12:27:46Z,direct_url=,disk_format='vmdk',id=82a9852c-0eb3-4602-b891-4c6c26b90c26,min_disk=1,min_ram=0,name='tempest-test-snap-2082815489',owner='0c78292171414530a1caba3dfd0f620f',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-08-29T12:28:02Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 853.706278] env[61728]: DEBUG nova.virt.hardware [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 853.706498] env[61728]: DEBUG nova.virt.hardware [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.706740] env[61728]: DEBUG nova.virt.hardware [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 853.706935] env[61728]: DEBUG nova.virt.hardware [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.708037] env[61728]: DEBUG nova.virt.hardware [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 853.708334] env[61728]: DEBUG nova.virt.hardware [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 853.708735] env[61728]: DEBUG nova.virt.hardware [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 853.708979] env[61728]: DEBUG nova.virt.hardware [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 853.709319] env[61728]: DEBUG nova.virt.hardware [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 853.710734] env[61728]: DEBUG nova.virt.hardware [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 853.711677] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec9b309e-73b6-41e6-bdde-af9a8e52f193 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.722791] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8023f4-fd2a-4794-b488-4a2d0dbf2540 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.812021] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464343, 'name': Rename_Task, 'duration_secs': 0.150948} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.812367] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 853.812684] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3639e082-fedd-44b4-a8ee-7eabd64c72f6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.821436] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Waiting for the task: (returnval){ [ 853.821436] env[61728]: value = "task-464344" [ 853.821436] env[61728]: _type = "Task" [ 853.821436] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.830881] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464344, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.886234] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.886516] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.887881] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-632814ac-824f-4e33-87f0-f61b17387591 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.910083] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-404c804f-6e2f-4258-99bc-c824da12beb3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.940234] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Reconfiguring VM to detach interface {{(pid=61728) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 853.940759] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d00e4883-528d-4c49-9446-9957b5b1755b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.962104] env[61728]: DEBUG oslo_vmware.api [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 853.962104] env[61728]: value = "task-464345" [ 853.962104] env[61728]: _type = "Task" [ 853.962104] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.973853] env[61728]: DEBUG oslo_vmware.api [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464345, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.139029] env[61728]: DEBUG oslo_concurrency.lockutils [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.198911] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b7990485-b760-453f-ba48-734eedfc13f7 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "7b16fbc0-7f13-405f-b84e-e18de1ca7dd2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.618s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.216652] env[61728]: INFO nova.compute.manager [-] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Took 1.38 seconds to deallocate network for instance. [ 854.270649] env[61728]: DEBUG nova.network.neutron [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Successfully updated port: 12471ef1-0e45-46ee-9b2a-b5008289a370 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.334126] env[61728]: DEBUG oslo_vmware.api [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464344, 'name': PowerOnVM_Task, 'duration_secs': 0.497466} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.334455] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 854.334756] env[61728]: INFO nova.compute.manager [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Took 8.20 seconds to spawn the instance on the hypervisor. [ 854.334956] env[61728]: DEBUG nova.compute.manager [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 854.335773] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-011b468d-eccc-424e-9b7c-e063f8ebed2f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.405014] env[61728]: DEBUG nova.compute.manager [req-05128d8d-30aa-4e8e-822b-ef75c1599f58 req-c6cf8c63-3a3e-4d56-b686-933b4613bb70 service nova] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Received event network-vif-plugged-12471ef1-0e45-46ee-9b2a-b5008289a370 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 854.405352] env[61728]: DEBUG oslo_concurrency.lockutils [req-05128d8d-30aa-4e8e-822b-ef75c1599f58 req-c6cf8c63-3a3e-4d56-b686-933b4613bb70 service nova] Acquiring lock "d8657743-30f9-4f8f-aca8-d98f3c557d19-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.405880] env[61728]: DEBUG oslo_concurrency.lockutils [req-05128d8d-30aa-4e8e-822b-ef75c1599f58 req-c6cf8c63-3a3e-4d56-b686-933b4613bb70 service nova] Lock "d8657743-30f9-4f8f-aca8-d98f3c557d19-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.405880] env[61728]: DEBUG oslo_concurrency.lockutils [req-05128d8d-30aa-4e8e-822b-ef75c1599f58 req-c6cf8c63-3a3e-4d56-b686-933b4613bb70 service nova] Lock "d8657743-30f9-4f8f-aca8-d98f3c557d19-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.406116] env[61728]: DEBUG nova.compute.manager [req-05128d8d-30aa-4e8e-822b-ef75c1599f58 req-c6cf8c63-3a3e-4d56-b686-933b4613bb70 service nova] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] No waiting events found dispatching network-vif-plugged-12471ef1-0e45-46ee-9b2a-b5008289a370 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 854.406348] env[61728]: WARNING nova.compute.manager [req-05128d8d-30aa-4e8e-822b-ef75c1599f58 req-c6cf8c63-3a3e-4d56-b686-933b4613bb70 service nova] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Received unexpected event network-vif-plugged-12471ef1-0e45-46ee-9b2a-b5008289a370 for instance with vm_state building and task_state spawning. [ 854.475683] env[61728]: DEBUG oslo_vmware.api [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.580081] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9444338c-48ab-47d5-9b4c-7babf1b644f0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.588901] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcdcb06d-9f45-4ece-9e68-5176e824e79b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.621564] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-310bf3af-a791-4d63-9dc7-e6390ed76155 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.630353] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f10f7d-5120-41b5-b370-771a420e972a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.646997] env[61728]: DEBUG nova.compute.provider_tree [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.723663] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.780931] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "refresh_cache-d8657743-30f9-4f8f-aca8-d98f3c557d19" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.780931] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "refresh_cache-d8657743-30f9-4f8f-aca8-d98f3c557d19" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.782947] env[61728]: DEBUG nova.network.neutron [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 854.863637] env[61728]: INFO nova.compute.manager [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Took 32.85 seconds to build instance. [ 854.974195] env[61728]: DEBUG oslo_vmware.api [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.150416] env[61728]: DEBUG nova.scheduler.client.report [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.336369] env[61728]: DEBUG nova.network.neutron [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 855.368640] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8dbbc1d-9b37-4890-9e9b-dc64eee20b9a tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Lock "a9745dc8-6400-424c-bc10-1181ee128648" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.388s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.476635] env[61728]: DEBUG oslo_vmware.api [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.488375] env[61728]: DEBUG nova.compute.manager [req-3f5f7dca-4b7a-4915-94ea-2c3d4a7bca06 req-0befa70a-002f-4de1-97a4-1817734b4d17 service nova] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Received event network-vif-deleted-1d776ec6-69e5-4f30-8201-13c69916cca0 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 855.588808] env[61728]: DEBUG nova.network.neutron [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Updating instance_info_cache with network_info: [{"id": "12471ef1-0e45-46ee-9b2a-b5008289a370", "address": "fa:16:3e:84:18:c3", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12471ef1-0e", "ovs_interfaceid": "12471ef1-0e45-46ee-9b2a-b5008289a370", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.597561] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "f9f07464-9f43-43fd-8895-2673861747fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.597829] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "f9f07464-9f43-43fd-8895-2673861747fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.655750] env[61728]: DEBUG oslo_concurrency.lockutils [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.989s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.658140] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.412s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.659817] env[61728]: INFO nova.compute.claims [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.684921] env[61728]: INFO nova.scheduler.client.report [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Deleted allocations for instance 6d55bb01-7855-4d86-ac57-ecae15ea590f [ 855.959960] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Acquiring lock "a9745dc8-6400-424c-bc10-1181ee128648" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.959960] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Lock "a9745dc8-6400-424c-bc10-1181ee128648" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.959960] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Acquiring lock "a9745dc8-6400-424c-bc10-1181ee128648-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.959960] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Lock "a9745dc8-6400-424c-bc10-1181ee128648-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.959960] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Lock "a9745dc8-6400-424c-bc10-1181ee128648-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.963205] env[61728]: INFO nova.compute.manager [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Terminating instance [ 855.964635] env[61728]: DEBUG nova.compute.manager [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 855.964833] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 855.965821] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dfe45ca-320c-4a57-bd05-53010c288d97 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.982246] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 855.982558] env[61728]: DEBUG oslo_vmware.api [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.982795] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-66660be4-df5e-4b45-8c5c-9cbb3b284b4d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.990767] env[61728]: DEBUG oslo_vmware.api [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Waiting for the task: (returnval){ [ 855.990767] env[61728]: value = "task-464346" [ 855.990767] env[61728]: _type = "Task" [ 855.990767] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.010702] env[61728]: DEBUG oslo_vmware.api [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464346, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.092704] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "refresh_cache-d8657743-30f9-4f8f-aca8-d98f3c557d19" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.093243] env[61728]: DEBUG nova.compute.manager [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Instance network_info: |[{"id": "12471ef1-0e45-46ee-9b2a-b5008289a370", "address": "fa:16:3e:84:18:c3", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12471ef1-0e", "ovs_interfaceid": "12471ef1-0e45-46ee-9b2a-b5008289a370", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 856.094932] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:18:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '12471ef1-0e45-46ee-9b2a-b5008289a370', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.105740] env[61728]: DEBUG oslo.service.loopingcall [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.106349] env[61728]: DEBUG nova.compute.manager [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 856.109747] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 856.110524] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f7672ef9-f1e7-44f1-adc4-8de30cfdbf01 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.133727] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.133727] env[61728]: value = "task-464347" [ 856.133727] env[61728]: _type = "Task" [ 856.133727] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.144358] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464347, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.194188] env[61728]: DEBUG oslo_concurrency.lockutils [None req-904bb167-645f-4a6c-aa7e-6479d9e2e51f tempest-ServerShowV254Test-614773774 tempest-ServerShowV254Test-614773774-project-member] Lock "6d55bb01-7855-4d86-ac57-ecae15ea590f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.285s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.479452] env[61728]: DEBUG oslo_vmware.api [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.501900] env[61728]: DEBUG oslo_vmware.api [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464346, 'name': PowerOffVM_Task, 'duration_secs': 0.17809} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.503418] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 856.503691] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 856.505067] env[61728]: DEBUG nova.compute.manager [req-d52153ec-8774-468d-a6cb-b0950bdeb144 req-e991962b-b187-48b3-9b35-130c4634290a service nova] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Received event network-changed-12471ef1-0e45-46ee-9b2a-b5008289a370 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 856.505608] env[61728]: DEBUG nova.compute.manager [req-d52153ec-8774-468d-a6cb-b0950bdeb144 req-e991962b-b187-48b3-9b35-130c4634290a service nova] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Refreshing instance network info cache due to event network-changed-12471ef1-0e45-46ee-9b2a-b5008289a370. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 856.506107] env[61728]: DEBUG oslo_concurrency.lockutils [req-d52153ec-8774-468d-a6cb-b0950bdeb144 req-e991962b-b187-48b3-9b35-130c4634290a service nova] Acquiring lock "refresh_cache-d8657743-30f9-4f8f-aca8-d98f3c557d19" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.506329] env[61728]: DEBUG oslo_concurrency.lockutils [req-d52153ec-8774-468d-a6cb-b0950bdeb144 req-e991962b-b187-48b3-9b35-130c4634290a service nova] Acquired lock "refresh_cache-d8657743-30f9-4f8f-aca8-d98f3c557d19" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.506552] env[61728]: DEBUG nova.network.neutron [req-d52153ec-8774-468d-a6cb-b0950bdeb144 req-e991962b-b187-48b3-9b35-130c4634290a service nova] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Refreshing network info cache for port 12471ef1-0e45-46ee-9b2a-b5008289a370 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 856.507736] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10f3c02b-6843-4af2-af34-f1a677257e38 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.614664] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 856.614664] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 856.614664] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Deleting the datastore file [datastore1] a9745dc8-6400-424c-bc10-1181ee128648 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 856.619814] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c6ca2a1-a8cc-4592-8b9b-44c1caced531 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.630507] env[61728]: DEBUG oslo_vmware.api [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Waiting for the task: (returnval){ [ 856.630507] env[61728]: value = "task-464349" [ 856.630507] env[61728]: _type = "Task" [ 856.630507] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.644865] env[61728]: DEBUG oslo_vmware.api [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464349, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.649018] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464347, 'name': CreateVM_Task, 'duration_secs': 0.424522} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.649018] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 856.650431] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.653019] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/82a9852c-0eb3-4602-b891-4c6c26b90c26" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.653019] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/82a9852c-0eb3-4602-b891-4c6c26b90c26" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.653019] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/82a9852c-0eb3-4602-b891-4c6c26b90c26" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.653019] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5681720-ad51-427b-b370-d8b4e9343b30 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.659213] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 856.659213] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb54a8-ba6e-27a2-6bf9-3b901e064be2" [ 856.659213] env[61728]: _type = "Task" [ 856.659213] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.666655] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb54a8-ba6e-27a2-6bf9-3b901e064be2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.982309] env[61728]: DEBUG oslo_vmware.api [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.052151] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fc91500-6bea-48c5-b90e-d87490d2c8e4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.061538] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fde5c1a-0143-4cde-91ac-f205129cc25a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.098925] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25de9b0f-6d0a-483d-8d15-3d643dec540a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.108286] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395db276-0cc7-43f6-882f-97833f5efd92 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.127329] env[61728]: DEBUG nova.compute.provider_tree [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.141028] env[61728]: DEBUG oslo_vmware.api [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Task: {'id': task-464349, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.31061} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.141096] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 857.141298] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 857.141486] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 857.141706] env[61728]: INFO nova.compute.manager [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Took 1.18 seconds to destroy the instance on the hypervisor. [ 857.141954] env[61728]: DEBUG oslo.service.loopingcall [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.142837] env[61728]: DEBUG nova.compute.manager [-] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 857.142936] env[61728]: DEBUG nova.network.neutron [-] [instance: a9745dc8-6400-424c-bc10-1181ee128648] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 857.171375] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/82a9852c-0eb3-4602-b891-4c6c26b90c26" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.171687] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Processing image 82a9852c-0eb3-4602-b891-4c6c26b90c26 {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.172203] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/82a9852c-0eb3-4602-b891-4c6c26b90c26/82a9852c-0eb3-4602-b891-4c6c26b90c26.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.172371] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/82a9852c-0eb3-4602-b891-4c6c26b90c26/82a9852c-0eb3-4602-b891-4c6c26b90c26.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.172564] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.172844] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-560146b8-9d5e-4671-938c-d73866d367f9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.183974] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.184250] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 857.185101] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4aaae97-e509-4055-ba7a-26cf69684f09 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.194224] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 857.194224] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520ba42d-32a6-e77f-4ad8-f493143e6a04" [ 857.194224] env[61728]: _type = "Task" [ 857.194224] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.201394] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520ba42d-32a6-e77f-4ad8-f493143e6a04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.344408] env[61728]: DEBUG nova.network.neutron [req-d52153ec-8774-468d-a6cb-b0950bdeb144 req-e991962b-b187-48b3-9b35-130c4634290a service nova] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Updated VIF entry in instance network info cache for port 12471ef1-0e45-46ee-9b2a-b5008289a370. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 857.345046] env[61728]: DEBUG nova.network.neutron [req-d52153ec-8774-468d-a6cb-b0950bdeb144 req-e991962b-b187-48b3-9b35-130c4634290a service nova] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Updating instance_info_cache with network_info: [{"id": "12471ef1-0e45-46ee-9b2a-b5008289a370", "address": "fa:16:3e:84:18:c3", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12471ef1-0e", "ovs_interfaceid": "12471ef1-0e45-46ee-9b2a-b5008289a370", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.480182] env[61728]: DEBUG oslo_vmware.api [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.632601] env[61728]: DEBUG nova.scheduler.client.report [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.671253] env[61728]: DEBUG nova.compute.manager [req-df714ac8-b08f-41c0-a075-23da79eeca39 req-3b8ddbe6-8bf6-40d0-b692-eec3e93df081 service nova] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Received event network-vif-deleted-3bd62766-6bca-43f7-b2a6-fb3f63eee1c7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 857.671253] env[61728]: INFO nova.compute.manager [req-df714ac8-b08f-41c0-a075-23da79eeca39 req-3b8ddbe6-8bf6-40d0-b692-eec3e93df081 service nova] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Neutron deleted interface 3bd62766-6bca-43f7-b2a6-fb3f63eee1c7; detaching it from the instance and deleting it from the info cache [ 857.671567] env[61728]: DEBUG nova.network.neutron [req-df714ac8-b08f-41c0-a075-23da79eeca39 req-3b8ddbe6-8bf6-40d0-b692-eec3e93df081 service nova] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.704283] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Preparing fetch location {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 857.704532] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Fetch image to [datastore1] OSTACK_IMG_db3588da-7ed5-4fd0-be7b-647bf1f49992/OSTACK_IMG_db3588da-7ed5-4fd0-be7b-647bf1f49992.vmdk {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 857.704739] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Downloading stream optimized image 82a9852c-0eb3-4602-b891-4c6c26b90c26 to [datastore1] OSTACK_IMG_db3588da-7ed5-4fd0-be7b-647bf1f49992/OSTACK_IMG_db3588da-7ed5-4fd0-be7b-647bf1f49992.vmdk on the data store datastore1 as vApp {{(pid=61728) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 857.704970] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Downloading image file data 82a9852c-0eb3-4602-b891-4c6c26b90c26 to the ESX as VM named 'OSTACK_IMG_db3588da-7ed5-4fd0-be7b-647bf1f49992' {{(pid=61728) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 857.798730] env[61728]: DEBUG oslo_vmware.rw_handles [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 857.798730] env[61728]: value = "resgroup-9" [ 857.798730] env[61728]: _type = "ResourcePool" [ 857.798730] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 857.798730] env[61728]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-f9024c79-bc9d-42d3-8bd5-32df006bcebf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.827247] env[61728]: DEBUG oslo_vmware.rw_handles [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lease: (returnval){ [ 857.827247] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529cf653-ff5a-41f5-60a8-4c066ba07d7d" [ 857.827247] env[61728]: _type = "HttpNfcLease" [ 857.827247] env[61728]: } obtained for vApp import into resource pool (val){ [ 857.827247] env[61728]: value = "resgroup-9" [ 857.827247] env[61728]: _type = "ResourcePool" [ 857.827247] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 857.827514] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the lease: (returnval){ [ 857.827514] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529cf653-ff5a-41f5-60a8-4c066ba07d7d" [ 857.827514] env[61728]: _type = "HttpNfcLease" [ 857.827514] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 857.838775] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 857.838775] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529cf653-ff5a-41f5-60a8-4c066ba07d7d" [ 857.838775] env[61728]: _type = "HttpNfcLease" [ 857.838775] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 857.851605] env[61728]: DEBUG oslo_concurrency.lockutils [req-d52153ec-8774-468d-a6cb-b0950bdeb144 req-e991962b-b187-48b3-9b35-130c4634290a service nova] Releasing lock "refresh_cache-d8657743-30f9-4f8f-aca8-d98f3c557d19" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.980828] env[61728]: DEBUG oslo_vmware.api [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.052544] env[61728]: DEBUG nova.network.neutron [-] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.139790] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.140386] env[61728]: DEBUG nova.compute.manager [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 858.147294] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.460s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.147294] env[61728]: DEBUG nova.objects.instance [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lazy-loading 'resources' on Instance uuid 2976d71f-7322-4640-b734-645dbb70e0c3 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.175125] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b6fe4d5d-5f2d-4005-bc68-e14720536bed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.187233] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e028305-29a3-4cdc-b90c-31a9a5e660cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.226111] env[61728]: DEBUG nova.compute.manager [req-df714ac8-b08f-41c0-a075-23da79eeca39 req-3b8ddbe6-8bf6-40d0-b692-eec3e93df081 service nova] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Detach interface failed, port_id=3bd62766-6bca-43f7-b2a6-fb3f63eee1c7, reason: Instance a9745dc8-6400-424c-bc10-1181ee128648 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 858.341190] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 858.341190] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529cf653-ff5a-41f5-60a8-4c066ba07d7d" [ 858.341190] env[61728]: _type = "HttpNfcLease" [ 858.341190] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 858.341190] env[61728]: DEBUG oslo_vmware.rw_handles [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 858.341190] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529cf653-ff5a-41f5-60a8-4c066ba07d7d" [ 858.341190] env[61728]: _type = "HttpNfcLease" [ 858.341190] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 858.341190] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b399a852-5ae7-442c-a0e4-a077745e51ab {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.356318] env[61728]: DEBUG oslo_vmware.rw_handles [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f1ae52-c4f7-b4ba-91b3-6cb74b2a3c6e/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 858.356582] env[61728]: DEBUG oslo_vmware.rw_handles [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f1ae52-c4f7-b4ba-91b3-6cb74b2a3c6e/disk-0.vmdk. {{(pid=61728) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 858.423623] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-840620e8-79ce-4eee-b4fb-2485beaa585b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.482282] env[61728]: DEBUG oslo_vmware.api [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.556058] env[61728]: INFO nova.compute.manager [-] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Took 1.41 seconds to deallocate network for instance. [ 858.650633] env[61728]: DEBUG nova.compute.utils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 858.652424] env[61728]: DEBUG nova.objects.instance [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lazy-loading 'numa_topology' on Instance uuid 2976d71f-7322-4640-b734-645dbb70e0c3 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.655480] env[61728]: DEBUG nova.compute.manager [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 858.655480] env[61728]: DEBUG nova.network.neutron [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 858.732338] env[61728]: DEBUG nova.policy [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '70856f8255c940ac979244efb02c43b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '276942c8cf8a42729a541de096f69a0c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 858.985484] env[61728]: DEBUG oslo_vmware.api [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.065129] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.076626] env[61728]: DEBUG oslo_vmware.rw_handles [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Completed reading data from the image iterator. {{(pid=61728) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 859.077073] env[61728]: DEBUG oslo_vmware.rw_handles [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f1ae52-c4f7-b4ba-91b3-6cb74b2a3c6e/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 859.078110] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2f0376-ee85-41c8-a7cb-f821de52128b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.102969] env[61728]: DEBUG oslo_vmware.rw_handles [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f1ae52-c4f7-b4ba-91b3-6cb74b2a3c6e/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 859.103282] env[61728]: DEBUG oslo_vmware.rw_handles [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f1ae52-c4f7-b4ba-91b3-6cb74b2a3c6e/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 859.103637] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-75529552-0708-49b1-9c12-1e298ada7314 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.154685] env[61728]: DEBUG nova.compute.manager [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 859.158042] env[61728]: DEBUG nova.objects.base [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Object Instance<2976d71f-7322-4640-b734-645dbb70e0c3> lazy-loaded attributes: resources,numa_topology {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 859.312810] env[61728]: DEBUG oslo_vmware.rw_handles [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f1ae52-c4f7-b4ba-91b3-6cb74b2a3c6e/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 859.313096] env[61728]: INFO nova.virt.vmwareapi.images [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Downloaded image file data 82a9852c-0eb3-4602-b891-4c6c26b90c26 [ 859.314709] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1717acb4-2e59-482e-ae77-fe3f6a1c1aa3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.335618] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-473965f1-bc1f-4618-a090-33a5ada144d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.365552] env[61728]: INFO nova.virt.vmwareapi.images [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] The imported VM was unregistered [ 859.368236] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Caching image {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 859.368519] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Creating directory with path [datastore1] devstack-image-cache_base/82a9852c-0eb3-4602-b891-4c6c26b90c26 {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.368808] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76851d4d-f644-42c5-832d-e628f0b96d62 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.392667] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Created directory with path [datastore1] devstack-image-cache_base/82a9852c-0eb3-4602-b891-4c6c26b90c26 {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.393091] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_db3588da-7ed5-4fd0-be7b-647bf1f49992/OSTACK_IMG_db3588da-7ed5-4fd0-be7b-647bf1f49992.vmdk to [datastore1] devstack-image-cache_base/82a9852c-0eb3-4602-b891-4c6c26b90c26/82a9852c-0eb3-4602-b891-4c6c26b90c26.vmdk. {{(pid=61728) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 859.393201] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-e06f13cb-d4bf-400b-a9f0-1f0432e66a8f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.401462] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 859.401462] env[61728]: value = "task-464352" [ 859.401462] env[61728]: _type = "Task" [ 859.401462] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.407845] env[61728]: DEBUG nova.network.neutron [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Successfully created port: 65322117-bd36-4d0d-b18c-dae8aac2fbba {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 859.415932] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464352, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.481938] env[61728]: DEBUG oslo_vmware.api [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464345, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.546904] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acd99ac-21ec-484f-801e-4d37c8b55cdc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.558028] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e00c2199-3be0-4edc-9c9d-5208b103e4ec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.598971] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-146165f0-5c68-4c1f-8fa4-658fe82dea9b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.611229] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cce7571-f7f3-4bad-b3dc-58dc1e78a120 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.629077] env[61728]: DEBUG nova.compute.provider_tree [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 859.918677] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464352, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.987342] env[61728]: DEBUG oslo_vmware.api [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464345, 'name': ReconfigVM_Task, 'duration_secs': 5.943859} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.987592] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.987810] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Reconfigured VM to detach interface {{(pid=61728) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 860.160302] env[61728]: ERROR nova.scheduler.client.report [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [req-2dd5f0d8-7cb6-405e-ae03-6c2e8a2efdc4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2dd5f0d8-7cb6-405e-ae03-6c2e8a2efdc4"}]} [ 860.169221] env[61728]: DEBUG nova.compute.manager [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 860.190822] env[61728]: DEBUG nova.scheduler.client.report [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 860.203349] env[61728]: DEBUG nova.virt.hardware [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 860.203609] env[61728]: DEBUG nova.virt.hardware [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 860.203772] env[61728]: DEBUG nova.virt.hardware [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 860.206514] env[61728]: DEBUG nova.virt.hardware [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 860.207389] env[61728]: DEBUG nova.virt.hardware [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 860.207389] env[61728]: DEBUG nova.virt.hardware [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 860.207389] env[61728]: DEBUG nova.virt.hardware [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 860.208416] env[61728]: DEBUG nova.virt.hardware [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 860.208416] env[61728]: DEBUG nova.virt.hardware [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 860.208416] env[61728]: DEBUG nova.virt.hardware [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 860.208416] env[61728]: DEBUG nova.virt.hardware [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 860.209191] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a43c1c8a-37e2-4a26-84b9-3a4f972b432a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.216431] env[61728]: DEBUG nova.scheduler.client.report [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 860.216711] env[61728]: DEBUG nova.compute.provider_tree [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 860.225608] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb533424-dddd-4390-9c93-c67afc612f2e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.234409] env[61728]: DEBUG nova.scheduler.client.report [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 860.237098] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.237348] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.261571] env[61728]: DEBUG nova.scheduler.client.report [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 860.415776] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464352, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.484752] env[61728]: DEBUG nova.compute.manager [req-0053564b-7fd8-4315-ae74-f4f72c6bfac8 req-7f073055-6e5a-447b-9d83-54ac19db2e2d service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Received event network-vif-deleted-02293cc2-3b2d-4b93-b27d-0e61b2780e89 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 860.485271] env[61728]: INFO nova.compute.manager [req-0053564b-7fd8-4315-ae74-f4f72c6bfac8 req-7f073055-6e5a-447b-9d83-54ac19db2e2d service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Neutron deleted interface 02293cc2-3b2d-4b93-b27d-0e61b2780e89; detaching it from the instance and deleting it from the info cache [ 860.485357] env[61728]: DEBUG nova.network.neutron [req-0053564b-7fd8-4315-ae74-f4f72c6bfac8 req-7f073055-6e5a-447b-9d83-54ac19db2e2d service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Updating instance_info_cache with network_info: [{"id": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "address": "fa:16:3e:44:e0:e6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e719ade-f3", "ovs_interfaceid": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b32267f3-fc0e-4870-b60d-664d82e90ca7", "address": "fa:16:3e:99:4f:52", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb32267f3-fc", "ovs_interfaceid": "b32267f3-fc0e-4870-b60d-664d82e90ca7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.665666] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d30d7f-9e10-475e-b795-b3ece9fd7f2a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.677652] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56bb8047-35aa-4832-aa09-a1e486ec53ab {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.719020] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbdd21f7-ee7b-4f86-990f-2ef81bd83c9e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.731990] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d956bcf2-6611-40bd-8c6e-6246907536fb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.743065] env[61728]: DEBUG nova.compute.manager [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 860.758574] env[61728]: DEBUG nova.compute.provider_tree [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 860.918029] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464352, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.991804] env[61728]: DEBUG oslo_concurrency.lockutils [req-0053564b-7fd8-4315-ae74-f4f72c6bfac8 req-7f073055-6e5a-447b-9d83-54ac19db2e2d service nova] Acquiring lock "f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.992829] env[61728]: DEBUG oslo_concurrency.lockutils [req-0053564b-7fd8-4315-ae74-f4f72c6bfac8 req-7f073055-6e5a-447b-9d83-54ac19db2e2d service nova] Acquired lock "f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.993484] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da1d781-dcd9-4052-9727-fba14aeb53b5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.017129] env[61728]: DEBUG oslo_concurrency.lockutils [req-0053564b-7fd8-4315-ae74-f4f72c6bfac8 req-7f073055-6e5a-447b-9d83-54ac19db2e2d service nova] Releasing lock "f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.017944] env[61728]: WARNING nova.compute.manager [req-0053564b-7fd8-4315-ae74-f4f72c6bfac8 req-7f073055-6e5a-447b-9d83-54ac19db2e2d service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Detach interface failed, port_id=02293cc2-3b2d-4b93-b27d-0e61b2780e89, reason: No device with interface-id 02293cc2-3b2d-4b93-b27d-0e61b2780e89 exists on VM: nova.exception.NotFound: No device with interface-id 02293cc2-3b2d-4b93-b27d-0e61b2780e89 exists on VM [ 861.244062] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "f28348d2-c062-497a-b374-521df51054ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.244062] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "f28348d2-c062-497a-b374-521df51054ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.244062] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "f28348d2-c062-497a-b374-521df51054ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.244062] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "f28348d2-c062-497a-b374-521df51054ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.244062] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "f28348d2-c062-497a-b374-521df51054ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.247836] env[61728]: INFO nova.compute.manager [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Terminating instance [ 861.251459] env[61728]: DEBUG nova.compute.manager [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 861.252159] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 861.254067] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6c11cb-a963-4c4e-856e-6079579da8d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.278985] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 861.280431] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8cea1593-a14f-4b35-9898-bf7e06550d37 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.295300] env[61728]: DEBUG oslo_vmware.api [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 861.295300] env[61728]: value = "task-464353" [ 861.295300] env[61728]: _type = "Task" [ 861.295300] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.296480] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.312155] env[61728]: DEBUG oslo_vmware.api [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464353, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.318218] env[61728]: DEBUG nova.scheduler.client.report [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 118 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 861.318218] env[61728]: DEBUG nova.compute.provider_tree [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 118 to 119 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 861.318218] env[61728]: DEBUG nova.compute.provider_tree [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 861.321452] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.321851] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.322241] env[61728]: DEBUG nova.network.neutron [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 861.417862] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464352, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.778656] env[61728]: DEBUG nova.network.neutron [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Successfully updated port: 65322117-bd36-4d0d-b18c-dae8aac2fbba {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 861.805891] env[61728]: DEBUG nova.compute.manager [req-223aa0f5-4114-487f-9724-a86176679fae req-759870ca-7b9f-4bcd-8c5f-5e2a2bc69a78 service nova] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Received event network-vif-plugged-65322117-bd36-4d0d-b18c-dae8aac2fbba {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 861.806144] env[61728]: DEBUG oslo_concurrency.lockutils [req-223aa0f5-4114-487f-9724-a86176679fae req-759870ca-7b9f-4bcd-8c5f-5e2a2bc69a78 service nova] Acquiring lock "87540771-c864-4d49-8ad3-b6b559841761-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.806356] env[61728]: DEBUG oslo_concurrency.lockutils [req-223aa0f5-4114-487f-9724-a86176679fae req-759870ca-7b9f-4bcd-8c5f-5e2a2bc69a78 service nova] Lock "87540771-c864-4d49-8ad3-b6b559841761-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.806524] env[61728]: DEBUG oslo_concurrency.lockutils [req-223aa0f5-4114-487f-9724-a86176679fae req-759870ca-7b9f-4bcd-8c5f-5e2a2bc69a78 service nova] Lock "87540771-c864-4d49-8ad3-b6b559841761-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.806696] env[61728]: DEBUG nova.compute.manager [req-223aa0f5-4114-487f-9724-a86176679fae req-759870ca-7b9f-4bcd-8c5f-5e2a2bc69a78 service nova] [instance: 87540771-c864-4d49-8ad3-b6b559841761] No waiting events found dispatching network-vif-plugged-65322117-bd36-4d0d-b18c-dae8aac2fbba {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 861.806898] env[61728]: WARNING nova.compute.manager [req-223aa0f5-4114-487f-9724-a86176679fae req-759870ca-7b9f-4bcd-8c5f-5e2a2bc69a78 service nova] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Received unexpected event network-vif-plugged-65322117-bd36-4d0d-b18c-dae8aac2fbba for instance with vm_state building and task_state spawning. [ 861.811012] env[61728]: DEBUG oslo_vmware.api [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464353, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.825417] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.680s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.829310] env[61728]: DEBUG oslo_concurrency.lockutils [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.107s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.829310] env[61728]: DEBUG nova.objects.instance [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Lazy-loading 'resources' on Instance uuid 4e5dc620-68a1-4e10-a8be-702c4999ca10 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.921411] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464352, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.486049} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.921785] env[61728]: INFO nova.virt.vmwareapi.ds_util [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_db3588da-7ed5-4fd0-be7b-647bf1f49992/OSTACK_IMG_db3588da-7ed5-4fd0-be7b-647bf1f49992.vmdk to [datastore1] devstack-image-cache_base/82a9852c-0eb3-4602-b891-4c6c26b90c26/82a9852c-0eb3-4602-b891-4c6c26b90c26.vmdk. [ 861.922028] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Cleaning up location [datastore1] OSTACK_IMG_db3588da-7ed5-4fd0-be7b-647bf1f49992 {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 861.922211] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_db3588da-7ed5-4fd0-be7b-647bf1f49992 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 861.922508] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-46ac60aa-88a4-42d8-a8bf-c5b0d123e756 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.929839] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 861.929839] env[61728]: value = "task-464354" [ 861.929839] env[61728]: _type = "Task" [ 861.929839] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.939678] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464354, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.278400] env[61728]: INFO nova.network.neutron [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Port b32267f3-fc0e-4870-b60d-664d82e90ca7 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 862.278779] env[61728]: DEBUG nova.network.neutron [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Updating instance_info_cache with network_info: [{"id": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "address": "fa:16:3e:44:e0:e6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e719ade-f3", "ovs_interfaceid": "0e719ade-f327-4ddb-8db5-8beda823ed1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.284495] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "refresh_cache-87540771-c864-4d49-8ad3-b6b559841761" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.284495] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired lock "refresh_cache-87540771-c864-4d49-8ad3-b6b559841761" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.284495] env[61728]: DEBUG nova.network.neutron [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 862.308540] env[61728]: DEBUG oslo_vmware.api [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464353, 'name': PowerOffVM_Task, 'duration_secs': 0.638666} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.308827] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 862.309009] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 862.309282] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc855051-f9bd-4c0b-944c-f18e1bee9c84 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.340028] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1dc0de96-4f07-472d-a8f2-69b2f73bb1a2 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "2976d71f-7322-4640-b734-645dbb70e0c3" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 48.758s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.341175] env[61728]: DEBUG oslo_concurrency.lockutils [None req-935810cb-9d43-4eff-b7bf-b8cb386237f1 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "2976d71f-7322-4640-b734-645dbb70e0c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 26.589s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.341497] env[61728]: DEBUG oslo_concurrency.lockutils [None req-935810cb-9d43-4eff-b7bf-b8cb386237f1 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "2976d71f-7322-4640-b734-645dbb70e0c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.341840] env[61728]: DEBUG oslo_concurrency.lockutils [None req-935810cb-9d43-4eff-b7bf-b8cb386237f1 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "2976d71f-7322-4640-b734-645dbb70e0c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.342100] env[61728]: DEBUG oslo_concurrency.lockutils [None req-935810cb-9d43-4eff-b7bf-b8cb386237f1 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "2976d71f-7322-4640-b734-645dbb70e0c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.344786] env[61728]: INFO nova.compute.manager [None req-935810cb-9d43-4eff-b7bf-b8cb386237f1 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Terminating instance [ 862.346813] env[61728]: DEBUG nova.compute.manager [None req-935810cb-9d43-4eff-b7bf-b8cb386237f1 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 862.347055] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-935810cb-9d43-4eff-b7bf-b8cb386237f1 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 862.347482] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3e41ad2c-7bcd-428e-bafa-766cb8cb97a7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.358393] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70338de2-adf8-4e89-9f20-2ffeac233fbb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.392836] env[61728]: WARNING nova.virt.vmwareapi.vmops [None req-935810cb-9d43-4eff-b7bf-b8cb386237f1 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2976d71f-7322-4640-b734-645dbb70e0c3 could not be found. [ 862.393105] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-935810cb-9d43-4eff-b7bf-b8cb386237f1 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 862.393300] env[61728]: INFO nova.compute.manager [None req-935810cb-9d43-4eff-b7bf-b8cb386237f1 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Took 0.05 seconds to destroy the instance on the hypervisor. [ 862.393550] env[61728]: DEBUG oslo.service.loopingcall [None req-935810cb-9d43-4eff-b7bf-b8cb386237f1 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.395985] env[61728]: DEBUG nova.compute.manager [-] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 862.396094] env[61728]: DEBUG nova.network.neutron [-] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 862.442428] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464354, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.042456} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.442696] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 862.443228] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/82a9852c-0eb3-4602-b891-4c6c26b90c26/82a9852c-0eb3-4602-b891-4c6c26b90c26.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.443228] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/82a9852c-0eb3-4602-b891-4c6c26b90c26/82a9852c-0eb3-4602-b891-4c6c26b90c26.vmdk to [datastore1] d8657743-30f9-4f8f-aca8-d98f3c557d19/d8657743-30f9-4f8f-aca8-d98f3c557d19.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 862.443355] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e246eb2-e7c6-4554-91df-49f99dc995e8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.451139] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 862.451139] env[61728]: value = "task-464356" [ 862.451139] env[61728]: _type = "Task" [ 862.451139] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.460100] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464356, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.491041] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 862.491301] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 862.491482] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Deleting the datastore file [datastore1] f28348d2-c062-497a-b374-521df51054ee {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 862.491833] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c5581f3-5e4a-4f52-b391-97dd60b46975 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.498948] env[61728]: DEBUG oslo_vmware.api [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 862.498948] env[61728]: value = "task-464357" [ 862.498948] env[61728]: _type = "Task" [ 862.498948] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.510539] env[61728]: DEBUG oslo_vmware.api [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464357, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.515070] env[61728]: DEBUG nova.compute.manager [req-0899887c-f95f-46c8-ba11-812b1d3ce235 req-2790c159-0afe-406e-84b4-291205f9e059 service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Received event network-vif-deleted-b32267f3-fc0e-4870-b60d-664d82e90ca7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 862.748454] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b57bce67-46e7-4ca4-810f-3f176fa0c5d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.758123] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51bd804e-2b15-4a8a-a237-a5dd6322d269 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.791395] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "refresh_cache-f28348d2-c062-497a-b374-521df51054ee" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.796578] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8319bc2a-83c7-4592-802f-a9e3a099900a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.806417] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-065af8d4-8254-4a7f-9d28-c9d17182727b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.822241] env[61728]: DEBUG nova.compute.provider_tree [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 862.870485] env[61728]: DEBUG nova.network.neutron [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 862.968395] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464356, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.012601] env[61728]: DEBUG oslo_vmware.api [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464357, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.173919] env[61728]: DEBUG nova.network.neutron [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Updating instance_info_cache with network_info: [{"id": "65322117-bd36-4d0d-b18c-dae8aac2fbba", "address": "fa:16:3e:75:41:c4", "network": {"id": "22ad4b68-2895-4651-803f-44b8031a6292", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-489227259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "276942c8cf8a42729a541de096f69a0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65322117-bd", "ovs_interfaceid": "65322117-bd36-4d0d-b18c-dae8aac2fbba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.259011] env[61728]: DEBUG nova.network.neutron [-] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.297385] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a29916f-e329-41c1-a5a4-07aae6db5ee0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-f28348d2-c062-497a-b374-521df51054ee-02293cc2-3b2d-4b93-b27d-0e61b2780e89" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.919s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.364033] env[61728]: DEBUG nova.scheduler.client.report [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 119 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 863.364033] env[61728]: DEBUG nova.compute.provider_tree [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 119 to 120 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 863.364033] env[61728]: DEBUG nova.compute.provider_tree [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 863.465651] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464356, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.516965] env[61728]: DEBUG oslo_vmware.api [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464357, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.849304} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.518054] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 863.518500] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 863.518500] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 863.518670] env[61728]: INFO nova.compute.manager [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: f28348d2-c062-497a-b374-521df51054ee] Took 2.27 seconds to destroy the instance on the hypervisor. [ 863.518884] env[61728]: DEBUG oslo.service.loopingcall [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.519112] env[61728]: DEBUG nova.compute.manager [-] [instance: f28348d2-c062-497a-b374-521df51054ee] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 863.519211] env[61728]: DEBUG nova.network.neutron [-] [instance: f28348d2-c062-497a-b374-521df51054ee] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 863.588524] env[61728]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port b32267f3-fc0e-4870-b60d-664d82e90ca7 could not be found.", "detail": ""}} {{(pid=61728) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 863.588781] env[61728]: DEBUG nova.network.neutron [-] Unable to show port b32267f3-fc0e-4870-b60d-664d82e90ca7 as it no longer exists. {{(pid=61728) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 863.676843] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Releasing lock "refresh_cache-87540771-c864-4d49-8ad3-b6b559841761" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.677321] env[61728]: DEBUG nova.compute.manager [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Instance network_info: |[{"id": "65322117-bd36-4d0d-b18c-dae8aac2fbba", "address": "fa:16:3e:75:41:c4", "network": {"id": "22ad4b68-2895-4651-803f-44b8031a6292", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-489227259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "276942c8cf8a42729a541de096f69a0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65322117-bd", "ovs_interfaceid": "65322117-bd36-4d0d-b18c-dae8aac2fbba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 863.677868] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:41:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0248a27a-1d7f-4195-987b-06bfc8467347', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '65322117-bd36-4d0d-b18c-dae8aac2fbba', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 863.687921] env[61728]: DEBUG oslo.service.loopingcall [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.688492] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 863.688845] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8bbb8340-00fc-44e2-b23f-3bcbdf0e12b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.715701] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 863.715701] env[61728]: value = "task-464358" [ 863.715701] env[61728]: _type = "Task" [ 863.715701] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.726796] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464358, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.762511] env[61728]: INFO nova.compute.manager [-] [instance: 2976d71f-7322-4640-b734-645dbb70e0c3] Took 1.37 seconds to deallocate network for instance. [ 863.869025] env[61728]: DEBUG oslo_concurrency.lockutils [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.041s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.873411] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.621s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.873411] env[61728]: DEBUG nova.objects.instance [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lazy-loading 'resources' on Instance uuid bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.901467] env[61728]: INFO nova.scheduler.client.report [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Deleted allocations for instance 4e5dc620-68a1-4e10-a8be-702c4999ca10 [ 863.967600] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464356, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.043255] env[61728]: DEBUG nova.compute.manager [req-4197908e-c617-4621-b110-4af003b42926 req-2e1897d1-a614-4632-8494-6c9f2d66e283 service nova] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Received event network-changed-65322117-bd36-4d0d-b18c-dae8aac2fbba {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 864.043466] env[61728]: DEBUG nova.compute.manager [req-4197908e-c617-4621-b110-4af003b42926 req-2e1897d1-a614-4632-8494-6c9f2d66e283 service nova] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Refreshing instance network info cache due to event network-changed-65322117-bd36-4d0d-b18c-dae8aac2fbba. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 864.043689] env[61728]: DEBUG oslo_concurrency.lockutils [req-4197908e-c617-4621-b110-4af003b42926 req-2e1897d1-a614-4632-8494-6c9f2d66e283 service nova] Acquiring lock "refresh_cache-87540771-c864-4d49-8ad3-b6b559841761" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.043857] env[61728]: DEBUG oslo_concurrency.lockutils [req-4197908e-c617-4621-b110-4af003b42926 req-2e1897d1-a614-4632-8494-6c9f2d66e283 service nova] Acquired lock "refresh_cache-87540771-c864-4d49-8ad3-b6b559841761" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.043968] env[61728]: DEBUG nova.network.neutron [req-4197908e-c617-4621-b110-4af003b42926 req-2e1897d1-a614-4632-8494-6c9f2d66e283 service nova] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Refreshing network info cache for port 65322117-bd36-4d0d-b18c-dae8aac2fbba {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 864.234181] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464358, 'name': CreateVM_Task, 'duration_secs': 0.460468} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.234378] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 864.235127] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.235313] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.235650] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 864.235919] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54fa8583-43f6-49b8-aadc-2d767d17a7b3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.254322] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 864.254322] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c0ff34-c213-9934-ee78-e01cfe64f32d" [ 864.254322] env[61728]: _type = "Task" [ 864.254322] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.280266] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c0ff34-c213-9934-ee78-e01cfe64f32d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.409202] env[61728]: DEBUG oslo_concurrency.lockutils [None req-af65a0e4-f926-43d3-8f82-50d7de5dff62 tempest-ServersV294TestFqdnHostnames-1288626545 tempest-ServersV294TestFqdnHostnames-1288626545-project-member] Lock "4e5dc620-68a1-4e10-a8be-702c4999ca10" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.246s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.468010] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464356, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.772813] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c0ff34-c213-9934-ee78-e01cfe64f32d, 'name': SearchDatastore_Task, 'duration_secs': 0.09107} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.777288] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.777607] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.777885] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.778193] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.778638] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.779036] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-858f5d13-1067-49f2-8193-513bef6812e1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.802443] env[61728]: DEBUG oslo_concurrency.lockutils [None req-935810cb-9d43-4eff-b7bf-b8cb386237f1 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "2976d71f-7322-4640-b734-645dbb70e0c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.461s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.806215] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.806431] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 864.807443] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3fa05e3-3f33-4034-b40e-bc7b5584b8fa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.816971] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 864.816971] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52415a57-1317-2a90-61eb-e50c0adc3f1f" [ 864.816971] env[61728]: _type = "Task" [ 864.816971] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.834155] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52415a57-1317-2a90-61eb-e50c0adc3f1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.844413] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c01632-86ba-4b7a-98bd-5bfd61ab330e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.855856] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0395e1-55f1-431e-a74d-bd47aee9b435 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.897700] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef6cdb5-a8bd-48f1-974e-c6a2df8c6de1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.912205] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ae783e-395e-42d6-be02-5ef7dafcc423 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.936970] env[61728]: DEBUG nova.network.neutron [-] [instance: f28348d2-c062-497a-b374-521df51054ee] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.939561] env[61728]: DEBUG nova.compute.provider_tree [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 864.970826] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464356, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.996313] env[61728]: DEBUG nova.network.neutron [req-4197908e-c617-4621-b110-4af003b42926 req-2e1897d1-a614-4632-8494-6c9f2d66e283 service nova] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Updated VIF entry in instance network info cache for port 65322117-bd36-4d0d-b18c-dae8aac2fbba. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 864.996313] env[61728]: DEBUG nova.network.neutron [req-4197908e-c617-4621-b110-4af003b42926 req-2e1897d1-a614-4632-8494-6c9f2d66e283 service nova] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Updating instance_info_cache with network_info: [{"id": "65322117-bd36-4d0d-b18c-dae8aac2fbba", "address": "fa:16:3e:75:41:c4", "network": {"id": "22ad4b68-2895-4651-803f-44b8031a6292", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-489227259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "276942c8cf8a42729a541de096f69a0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65322117-bd", "ovs_interfaceid": "65322117-bd36-4d0d-b18c-dae8aac2fbba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.333842] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52415a57-1317-2a90-61eb-e50c0adc3f1f, 'name': SearchDatastore_Task, 'duration_secs': 0.081643} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.333842] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0de5d925-597c-4a1d-b10b-849dc4954c0e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.339959] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 865.339959] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e5c28e-24f5-2cb4-ac03-d1a20cb75152" [ 865.339959] env[61728]: _type = "Task" [ 865.339959] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.348747] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e5c28e-24f5-2cb4-ac03-d1a20cb75152, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.439498] env[61728]: INFO nova.compute.manager [-] [instance: f28348d2-c062-497a-b374-521df51054ee] Took 1.92 seconds to deallocate network for instance. [ 865.468778] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464356, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.827457} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.470141] env[61728]: ERROR nova.scheduler.client.report [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] [req-1605fefe-94fe-43dc-84de-ca4915987627] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1605fefe-94fe-43dc-84de-ca4915987627"}]} [ 865.471443] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/82a9852c-0eb3-4602-b891-4c6c26b90c26/82a9852c-0eb3-4602-b891-4c6c26b90c26.vmdk to [datastore1] d8657743-30f9-4f8f-aca8-d98f3c557d19/d8657743-30f9-4f8f-aca8-d98f3c557d19.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 865.474263] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d85b14-5700-4f90-b433-6d3dfcfcbaf6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.505971] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] d8657743-30f9-4f8f-aca8-d98f3c557d19/d8657743-30f9-4f8f-aca8-d98f3c557d19.vmdk or device None with type streamOptimized {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.508432] env[61728]: DEBUG nova.scheduler.client.report [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 865.510892] env[61728]: DEBUG oslo_concurrency.lockutils [req-4197908e-c617-4621-b110-4af003b42926 req-2e1897d1-a614-4632-8494-6c9f2d66e283 service nova] Releasing lock "refresh_cache-87540771-c864-4d49-8ad3-b6b559841761" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.511252] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d04b02e-fcb4-4606-a104-39f9f16966fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.532646] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 865.532646] env[61728]: value = "task-464359" [ 865.532646] env[61728]: _type = "Task" [ 865.532646] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.542237] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464359, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.543303] env[61728]: DEBUG nova.scheduler.client.report [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 865.543519] env[61728]: DEBUG nova.compute.provider_tree [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 865.555918] env[61728]: DEBUG nova.scheduler.client.report [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 865.574479] env[61728]: DEBUG nova.scheduler.client.report [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 865.848394] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "ae25dc25-f600-471f-aaaa-f166ed691bf3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.849022] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "ae25dc25-f600-471f-aaaa-f166ed691bf3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.859466] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e5c28e-24f5-2cb4-ac03-d1a20cb75152, 'name': SearchDatastore_Task, 'duration_secs': 0.010453} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.860249] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.860249] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 87540771-c864-4d49-8ad3-b6b559841761/87540771-c864-4d49-8ad3-b6b559841761.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 865.860455] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1f66b214-fd3f-4c4d-805e-bfc7230056ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.868046] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 865.868046] env[61728]: value = "task-464360" [ 865.868046] env[61728]: _type = "Task" [ 865.868046] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.878564] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464360, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.955060] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.995575] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b193d600-6f89-4abc-aa99-7b99584fd69b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.005281] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e5a3531-92c2-4d72-9570-00a91fe0a6e3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.050645] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ffa0288-6075-473b-96b4-809c97be494c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.063397] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6682a88-4e26-4587-8d65-58ee3a08b87d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.068687] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464359, 'name': ReconfigVM_Task, 'duration_secs': 0.345013} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.069061] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Reconfigured VM instance instance-00000049 to attach disk [datastore1] d8657743-30f9-4f8f-aca8-d98f3c557d19/d8657743-30f9-4f8f-aca8-d98f3c557d19.vmdk or device None with type streamOptimized {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.070184] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-057e0a3c-95d7-4f0f-b9d3-6f64935273bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.081961] env[61728]: DEBUG nova.compute.provider_tree [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 866.084638] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 866.084638] env[61728]: value = "task-464361" [ 866.084638] env[61728]: _type = "Task" [ 866.084638] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.107774] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464361, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.143711] env[61728]: DEBUG nova.compute.manager [req-b2de2543-e770-4b1e-b48f-df427844fe46 req-6a16c9d6-5307-43b1-b87e-0b48444b188b service nova] [instance: f28348d2-c062-497a-b374-521df51054ee] Received event network-vif-deleted-0e719ade-f327-4ddb-8db5-8beda823ed1b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 866.383047] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464360, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.600150] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464361, 'name': Rename_Task, 'duration_secs': 0.216352} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.600464] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 866.600758] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf4e8ef2-901a-4a2e-b0cc-a27286fb60b5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.609082] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 866.609082] env[61728]: value = "task-464362" [ 866.609082] env[61728]: _type = "Task" [ 866.609082] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.617960] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464362, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.619170] env[61728]: DEBUG nova.scheduler.client.report [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 121 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 866.619419] env[61728]: DEBUG nova.compute.provider_tree [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 121 to 122 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 866.619636] env[61728]: DEBUG nova.compute.provider_tree [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 866.882135] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464360, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.762021} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.882416] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 87540771-c864-4d49-8ad3-b6b559841761/87540771-c864-4d49-8ad3-b6b559841761.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 866.882653] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.882899] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-400fa1a5-5b6c-4ebb-ad8c-1a9cf29a642c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.891510] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 866.891510] env[61728]: value = "task-464363" [ 866.891510] env[61728]: _type = "Task" [ 866.891510] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.902620] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464363, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.120922] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464362, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.124882] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.252s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.127426] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.466s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.127691] env[61728]: DEBUG nova.objects.instance [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lazy-loading 'resources' on Instance uuid 771341ed-8b8e-470c-9686-82650f5271b4 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.150335] env[61728]: INFO nova.scheduler.client.report [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Deleted allocations for instance bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5 [ 867.402543] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464363, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079152} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.402823] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.403703] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a4e554-53d9-4232-91ca-ebdf429406b2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.426409] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 87540771-c864-4d49-8ad3-b6b559841761/87540771-c864-4d49-8ad3-b6b559841761.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.426725] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb58877d-e701-4c32-a197-7bd323f1b5f6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.448643] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 867.448643] env[61728]: value = "task-464364" [ 867.448643] env[61728]: _type = "Task" [ 867.448643] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.457811] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464364, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.622564] env[61728]: DEBUG oslo_vmware.api [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464362, 'name': PowerOnVM_Task, 'duration_secs': 0.535076} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.622886] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 867.623115] env[61728]: INFO nova.compute.manager [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Took 13.95 seconds to spawn the instance on the hypervisor. [ 867.623304] env[61728]: DEBUG nova.compute.manager [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.624246] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce05ed58-ce87-47f6-a344-3eb3391e9ca0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.658357] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50354a02-c35a-47fd-b943-4ab90b63e8a9 tempest-ListImageFiltersTestJSON-795312603 tempest-ListImageFiltersTestJSON-795312603-project-member] Lock "bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.418s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.962090] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464364, 'name': ReconfigVM_Task, 'duration_secs': 0.320875} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.962411] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 87540771-c864-4d49-8ad3-b6b559841761/87540771-c864-4d49-8ad3-b6b559841761.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.963296] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-db3a2bc1-aa74-442a-aa3d-69d2704f3e13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.971851] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 867.971851] env[61728]: value = "task-464365" [ 867.971851] env[61728]: _type = "Task" [ 867.971851] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.983042] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464365, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.001181] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85e3da1-9e66-41c1-9e1f-1bd40dc2eea1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.009892] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85835844-cb16-4362-b3de-de3728e7356a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.049037] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6784968a-322b-49cf-bace-ba464cd4c54d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.058582] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb35d32c-8966-4071-ad3f-68aec202e038 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.071790] env[61728]: DEBUG nova.compute.provider_tree [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.148414] env[61728]: INFO nova.compute.manager [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Took 41.83 seconds to build instance. [ 868.484015] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464365, 'name': Rename_Task, 'duration_secs': 0.15551} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.484015] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 868.484015] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff510baa-4202-44cb-8829-2b12f2513842 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.491888] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 868.491888] env[61728]: value = "task-464366" [ 868.491888] env[61728]: _type = "Task" [ 868.491888] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.500844] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464366, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.575166] env[61728]: DEBUG nova.scheduler.client.report [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.649395] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ca1f62fd-3757-47c7-a23f-ccd163ee4fd9 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "d8657743-30f9-4f8f-aca8-d98f3c557d19" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.347s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.008208] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464366, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.082937] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.953s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.082937] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.273s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.085309] env[61728]: INFO nova.compute.claims [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.112697] env[61728]: INFO nova.scheduler.client.report [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Deleted allocations for instance 771341ed-8b8e-470c-9686-82650f5271b4 [ 869.151720] env[61728]: DEBUG nova.compute.manager [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 869.349049] env[61728]: DEBUG oslo_concurrency.lockutils [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "d8657743-30f9-4f8f-aca8-d98f3c557d19" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.349049] env[61728]: DEBUG oslo_concurrency.lockutils [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "d8657743-30f9-4f8f-aca8-d98f3c557d19" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.349049] env[61728]: DEBUG oslo_concurrency.lockutils [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "d8657743-30f9-4f8f-aca8-d98f3c557d19-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.349049] env[61728]: DEBUG oslo_concurrency.lockutils [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "d8657743-30f9-4f8f-aca8-d98f3c557d19-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.349049] env[61728]: DEBUG oslo_concurrency.lockutils [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "d8657743-30f9-4f8f-aca8-d98f3c557d19-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.349748] env[61728]: INFO nova.compute.manager [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Terminating instance [ 869.354472] env[61728]: DEBUG nova.compute.manager [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 869.354472] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 869.354472] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2a1ea2-e559-481b-a9af-1f041bac82d1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.363614] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 869.364092] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9bcf7715-7d76-4597-9597-a64cd1105824 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.371560] env[61728]: DEBUG oslo_vmware.api [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 869.371560] env[61728]: value = "task-464367" [ 869.371560] env[61728]: _type = "Task" [ 869.371560] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.380132] env[61728]: DEBUG oslo_vmware.api [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464367, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.504270] env[61728]: DEBUG oslo_vmware.api [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464366, 'name': PowerOnVM_Task, 'duration_secs': 0.706984} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.504668] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 869.504977] env[61728]: INFO nova.compute.manager [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Took 9.34 seconds to spawn the instance on the hypervisor. [ 869.505268] env[61728]: DEBUG nova.compute.manager [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.506354] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5052a5fe-f8f6-4d82-8780-c08271467fb1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.624711] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8c355af-0cb0-42b1-bb2d-0d75a8a0b438 tempest-ServersAdminTestJSON-1426658535 tempest-ServersAdminTestJSON-1426658535-project-member] Lock "771341ed-8b8e-470c-9686-82650f5271b4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.490s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.681697] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.886611] env[61728]: DEBUG oslo_vmware.api [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464367, 'name': PowerOffVM_Task, 'duration_secs': 0.221054} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.886611] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 869.886611] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 869.886611] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-308b5a6c-698e-4a0e-b206-a5b7dfce94c3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.029136] env[61728]: INFO nova.compute.manager [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Took 37.81 seconds to build instance. [ 870.195820] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Acquiring lock "2400abb1-8a07-4f6b-8818-778a2105beb8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.195820] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Lock "2400abb1-8a07-4f6b-8818-778a2105beb8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.336787] env[61728]: DEBUG oslo_concurrency.lockutils [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.337707] env[61728]: DEBUG oslo_concurrency.lockutils [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.337707] env[61728]: INFO nova.compute.manager [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Shelving [ 870.380168] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 870.380412] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 870.380722] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleting the datastore file [datastore1] d8657743-30f9-4f8f-aca8-d98f3c557d19 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 870.381015] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-42390aea-9209-439b-9b69-b1ef1cbd40b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.388761] env[61728]: DEBUG oslo_vmware.api [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 870.388761] env[61728]: value = "task-464369" [ 870.388761] env[61728]: _type = "Task" [ 870.388761] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.410579] env[61728]: DEBUG oslo_vmware.api [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464369, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.528333] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36e0e2d-b015-47c5-8d46-3aefb78436c0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.533026] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c67fe561-ab4d-4819-aff0-390e88684688 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "87540771-c864-4d49-8ad3-b6b559841761" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.339s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.540573] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55240430-9847-432d-8d06-4e0e21b38a9b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.583251] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abec1e4c-04f5-474a-beb8-e87dcab3c7bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.591556] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fe5c7b-2f45-4219-962c-ccb6da28b025 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.606704] env[61728]: DEBUG nova.compute.provider_tree [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.845792] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 870.846308] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9be1e80a-9776-4ad4-890e-b3e234ffd6a4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.856463] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 870.856463] env[61728]: value = "task-464370" [ 870.856463] env[61728]: _type = "Task" [ 870.856463] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.867735] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464370, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.904788] env[61728]: DEBUG oslo_vmware.api [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464369, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191845} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.905095] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.905661] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 870.905661] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 870.905819] env[61728]: INFO nova.compute.manager [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Took 1.55 seconds to destroy the instance on the hypervisor. [ 870.905922] env[61728]: DEBUG oslo.service.loopingcall [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.906141] env[61728]: DEBUG nova.compute.manager [-] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 870.906447] env[61728]: DEBUG nova.network.neutron [-] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 871.037109] env[61728]: DEBUG nova.compute.manager [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 871.110567] env[61728]: DEBUG nova.scheduler.client.report [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.208599] env[61728]: DEBUG nova.compute.manager [req-409ca73a-d155-44b0-bdfa-6e6ee6910ca7 req-f225d9ea-824d-4ad1-9d5d-cbca4a451fd5 service nova] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Received event network-changed-65322117-bd36-4d0d-b18c-dae8aac2fbba {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 871.208861] env[61728]: DEBUG nova.compute.manager [req-409ca73a-d155-44b0-bdfa-6e6ee6910ca7 req-f225d9ea-824d-4ad1-9d5d-cbca4a451fd5 service nova] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Refreshing instance network info cache due to event network-changed-65322117-bd36-4d0d-b18c-dae8aac2fbba. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 871.209424] env[61728]: DEBUG oslo_concurrency.lockutils [req-409ca73a-d155-44b0-bdfa-6e6ee6910ca7 req-f225d9ea-824d-4ad1-9d5d-cbca4a451fd5 service nova] Acquiring lock "refresh_cache-87540771-c864-4d49-8ad3-b6b559841761" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.209680] env[61728]: DEBUG oslo_concurrency.lockutils [req-409ca73a-d155-44b0-bdfa-6e6ee6910ca7 req-f225d9ea-824d-4ad1-9d5d-cbca4a451fd5 service nova] Acquired lock "refresh_cache-87540771-c864-4d49-8ad3-b6b559841761" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.209930] env[61728]: DEBUG nova.network.neutron [req-409ca73a-d155-44b0-bdfa-6e6ee6910ca7 req-f225d9ea-824d-4ad1-9d5d-cbca4a451fd5 service nova] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Refreshing network info cache for port 65322117-bd36-4d0d-b18c-dae8aac2fbba {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 871.368036] env[61728]: DEBUG nova.compute.manager [req-273f9076-e815-4d97-b50b-8ea03f80929c req-df8a18d3-f3d7-432f-bdbc-453d709c496e service nova] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Received event network-vif-deleted-12471ef1-0e45-46ee-9b2a-b5008289a370 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 871.368036] env[61728]: INFO nova.compute.manager [req-273f9076-e815-4d97-b50b-8ea03f80929c req-df8a18d3-f3d7-432f-bdbc-453d709c496e service nova] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Neutron deleted interface 12471ef1-0e45-46ee-9b2a-b5008289a370; detaching it from the instance and deleting it from the info cache [ 871.368180] env[61728]: DEBUG nova.network.neutron [req-273f9076-e815-4d97-b50b-8ea03f80929c req-df8a18d3-f3d7-432f-bdbc-453d709c496e service nova] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.374355] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464370, 'name': PowerOffVM_Task, 'duration_secs': 0.204404} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.374837] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 871.375792] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ea307f-04b3-49d4-b773-4dc4dcca2dab {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.398271] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d7b096-9886-4a9b-8535-e072792a294b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.567772] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.616452] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.533s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.617100] env[61728]: DEBUG nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 871.622212] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.788s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.622212] env[61728]: INFO nova.compute.claims [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 871.846970] env[61728]: DEBUG nova.network.neutron [-] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.873181] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-35ff435e-71aa-4ab6-a178-6f04e140046e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.888100] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8001f4ad-c3b8-4366-8366-10cff9d40abe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.911861] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 871.912285] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-76947681-ea57-472e-a9e9-9402df30b521 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.930725] env[61728]: DEBUG nova.compute.manager [req-273f9076-e815-4d97-b50b-8ea03f80929c req-df8a18d3-f3d7-432f-bdbc-453d709c496e service nova] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Detach interface failed, port_id=12471ef1-0e45-46ee-9b2a-b5008289a370, reason: Instance d8657743-30f9-4f8f-aca8-d98f3c557d19 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 871.931352] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 871.931352] env[61728]: value = "task-464371" [ 871.931352] env[61728]: _type = "Task" [ 871.931352] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.942681] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464371, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.126799] env[61728]: DEBUG nova.compute.utils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.131847] env[61728]: DEBUG nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 872.132055] env[61728]: DEBUG nova.network.neutron [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 872.137630] env[61728]: DEBUG nova.network.neutron [req-409ca73a-d155-44b0-bdfa-6e6ee6910ca7 req-f225d9ea-824d-4ad1-9d5d-cbca4a451fd5 service nova] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Updated VIF entry in instance network info cache for port 65322117-bd36-4d0d-b18c-dae8aac2fbba. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 872.138079] env[61728]: DEBUG nova.network.neutron [req-409ca73a-d155-44b0-bdfa-6e6ee6910ca7 req-f225d9ea-824d-4ad1-9d5d-cbca4a451fd5 service nova] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Updating instance_info_cache with network_info: [{"id": "65322117-bd36-4d0d-b18c-dae8aac2fbba", "address": "fa:16:3e:75:41:c4", "network": {"id": "22ad4b68-2895-4651-803f-44b8031a6292", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-489227259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "276942c8cf8a42729a541de096f69a0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65322117-bd", "ovs_interfaceid": "65322117-bd36-4d0d-b18c-dae8aac2fbba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.232244] env[61728]: DEBUG nova.policy [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8938cfb5439247d6b95adfe040babded', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9523aa656ff2459b94bb28d133fa4b05', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 872.352962] env[61728]: INFO nova.compute.manager [-] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Took 1.45 seconds to deallocate network for instance. [ 872.445218] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464371, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.637810] env[61728]: DEBUG nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 872.644578] env[61728]: DEBUG oslo_concurrency.lockutils [req-409ca73a-d155-44b0-bdfa-6e6ee6910ca7 req-f225d9ea-824d-4ad1-9d5d-cbca4a451fd5 service nova] Releasing lock "refresh_cache-87540771-c864-4d49-8ad3-b6b559841761" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.746580] env[61728]: DEBUG nova.network.neutron [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Successfully created port: 58d83a36-5bb8-4f09-bd16-16d01ec692d7 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 872.867993] env[61728]: DEBUG oslo_concurrency.lockutils [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.956146] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464371, 'name': CreateSnapshot_Task, 'duration_secs': 0.92713} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.959289] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 872.960447] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d657e8-027a-432b-a605-7a1ca0bdceb8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.069091] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca7a913-4d71-430f-b716-b9ac56391c42 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.078083] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64964dd8-29d5-4368-a853-66f1192ec593 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.115547] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982ca966-034a-4eff-baa0-0953f7a1399e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.125426] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2dfd47-2c57-49f9-84d3-99d50a29a4a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.150886] env[61728]: DEBUG nova.compute.provider_tree [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.482098] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 873.482098] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6c2b5103-6064-49d8-be59-42999c14e126 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.490859] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 873.490859] env[61728]: value = "task-464372" [ 873.490859] env[61728]: _type = "Task" [ 873.490859] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.500861] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464372, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.653557] env[61728]: DEBUG nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 873.657414] env[61728]: DEBUG nova.scheduler.client.report [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.687994] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 873.688317] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 873.688452] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 873.688661] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 873.688777] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 873.688943] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 873.689178] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 873.689342] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 873.689531] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 873.689704] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 873.689885] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.690857] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfed96bd-2548-4f09-ae0f-2986aee4817c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.700264] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e09cb90a-0cc9-49a6-aaf2-5ca4fa3664ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.003447] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464372, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.163122] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.163122] env[61728]: DEBUG nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 874.166624] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.051s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.166841] env[61728]: DEBUG nova.objects.instance [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lazy-loading 'resources' on Instance uuid 3ef9bf60-c6a1-4b7a-a375-5397fb871850 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.503765] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464372, 'name': CloneVM_Task} progress is 95%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.673045] env[61728]: DEBUG nova.compute.utils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 874.675406] env[61728]: DEBUG nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 874.675406] env[61728]: DEBUG nova.network.neutron [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 874.683723] env[61728]: DEBUG nova.network.neutron [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Successfully updated port: 58d83a36-5bb8-4f09-bd16-16d01ec692d7 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 874.746666] env[61728]: DEBUG nova.policy [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8938cfb5439247d6b95adfe040babded', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9523aa656ff2459b94bb28d133fa4b05', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 874.824503] env[61728]: DEBUG nova.compute.manager [req-0d4e3c38-43d6-486c-a095-37f80ef5a8ec req-ea5742a4-c345-4685-9dd4-2de434c3917c service nova] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Received event network-vif-plugged-58d83a36-5bb8-4f09-bd16-16d01ec692d7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 874.824669] env[61728]: DEBUG oslo_concurrency.lockutils [req-0d4e3c38-43d6-486c-a095-37f80ef5a8ec req-ea5742a4-c345-4685-9dd4-2de434c3917c service nova] Acquiring lock "121bef05-ef7b-47ee-b737-15f8b9d91158-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.825010] env[61728]: DEBUG oslo_concurrency.lockutils [req-0d4e3c38-43d6-486c-a095-37f80ef5a8ec req-ea5742a4-c345-4685-9dd4-2de434c3917c service nova] Lock "121bef05-ef7b-47ee-b737-15f8b9d91158-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.825754] env[61728]: DEBUG oslo_concurrency.lockutils [req-0d4e3c38-43d6-486c-a095-37f80ef5a8ec req-ea5742a4-c345-4685-9dd4-2de434c3917c service nova] Lock "121bef05-ef7b-47ee-b737-15f8b9d91158-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.825754] env[61728]: DEBUG nova.compute.manager [req-0d4e3c38-43d6-486c-a095-37f80ef5a8ec req-ea5742a4-c345-4685-9dd4-2de434c3917c service nova] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] No waiting events found dispatching network-vif-plugged-58d83a36-5bb8-4f09-bd16-16d01ec692d7 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 874.825754] env[61728]: WARNING nova.compute.manager [req-0d4e3c38-43d6-486c-a095-37f80ef5a8ec req-ea5742a4-c345-4685-9dd4-2de434c3917c service nova] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Received unexpected event network-vif-plugged-58d83a36-5bb8-4f09-bd16-16d01ec692d7 for instance with vm_state building and task_state spawning. [ 874.841759] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "0e77ef23-3d1a-4f00-83f9-5b617334f980" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.842080] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "0e77ef23-3d1a-4f00-83f9-5b617334f980" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.875527] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "2c166dd3-985c-4637-92ab-939b46a7c90e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.875527] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "2c166dd3-985c-4637-92ab-939b46a7c90e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.915643] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.915643] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.006596] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464372, 'name': CloneVM_Task, 'duration_secs': 1.267261} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.006897] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Created linked-clone VM from snapshot [ 875.009723] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c95c67-2c88-4d78-9611-6c7756621f77 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.018311] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Uploading image 6bf918f6-e123-44bd-a072-fe70029d1786 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 875.052586] env[61728]: DEBUG oslo_vmware.rw_handles [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 875.052586] env[61728]: value = "vm-122130" [ 875.052586] env[61728]: _type = "VirtualMachine" [ 875.052586] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 875.052964] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7c2c0c3b-8e0c-44eb-a417-cdd14796aa6b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.063216] env[61728]: DEBUG oslo_vmware.rw_handles [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lease: (returnval){ [ 875.063216] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529bbe7f-4c15-1f5d-2696-d017ef6c797c" [ 875.063216] env[61728]: _type = "HttpNfcLease" [ 875.063216] env[61728]: } obtained for exporting VM: (result){ [ 875.063216] env[61728]: value = "vm-122130" [ 875.063216] env[61728]: _type = "VirtualMachine" [ 875.063216] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 875.063484] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the lease: (returnval){ [ 875.063484] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529bbe7f-4c15-1f5d-2696-d017ef6c797c" [ 875.063484] env[61728]: _type = "HttpNfcLease" [ 875.063484] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 875.072240] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 875.072240] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529bbe7f-4c15-1f5d-2696-d017ef6c797c" [ 875.072240] env[61728]: _type = "HttpNfcLease" [ 875.072240] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 875.140835] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81dd4fdb-850a-4c9b-bfff-0e78aa37a254 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.145200] env[61728]: DEBUG nova.network.neutron [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Successfully created port: d16b0ff3-2873-4aa9-b5da-c97c6e5d0633 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 875.151413] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bcd218d-4f30-4469-b843-b5665f0c4030 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.185052] env[61728]: DEBUG nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 875.191197] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "refresh_cache-121bef05-ef7b-47ee-b737-15f8b9d91158" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.191197] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired lock "refresh_cache-121bef05-ef7b-47ee-b737-15f8b9d91158" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.191197] env[61728]: DEBUG nova.network.neutron [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 875.191197] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f9a03a-388c-45d5-adf8-21205c3a449e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.200423] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a5ea9f-50aa-407e-862e-ccc5666d33a4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.222613] env[61728]: DEBUG nova.compute.provider_tree [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.572668] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 875.572668] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529bbe7f-4c15-1f5d-2696-d017ef6c797c" [ 875.572668] env[61728]: _type = "HttpNfcLease" [ 875.572668] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 875.572960] env[61728]: DEBUG oslo_vmware.rw_handles [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 875.572960] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529bbe7f-4c15-1f5d-2696-d017ef6c797c" [ 875.572960] env[61728]: _type = "HttpNfcLease" [ 875.572960] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 875.574037] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8608c078-3b9f-4e99-bf86-f250986c755d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.581672] env[61728]: DEBUG oslo_vmware.rw_handles [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5284986e-eff9-c129-3129-76bc62dac282/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 875.581849] env[61728]: DEBUG oslo_vmware.rw_handles [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5284986e-eff9-c129-3129-76bc62dac282/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 875.727201] env[61728]: DEBUG nova.scheduler.client.report [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 875.734747] env[61728]: DEBUG nova.network.neutron [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 875.739964] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-362a4a0e-9bef-4bfd-a310-45b28dd47aa8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.884927] env[61728]: DEBUG nova.network.neutron [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Updating instance_info_cache with network_info: [{"id": "58d83a36-5bb8-4f09-bd16-16d01ec692d7", "address": "fa:16:3e:59:8b:54", "network": {"id": "74344aaa-7298-4ae0-ab9c-ebeef230b69d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1018009435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9523aa656ff2459b94bb28d133fa4b05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58d83a36-5b", "ovs_interfaceid": "58d83a36-5bb8-4f09-bd16-16d01ec692d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.194686] env[61728]: DEBUG nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 876.218171] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.218530] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.218766] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.219071] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.219266] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.219514] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.219774] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.219950] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.220142] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.220321] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.220780] env[61728]: DEBUG nova.virt.hardware [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.221532] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb03196-dfbb-4d41-8b07-32ac0dca0320 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.230994] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375ba947-9e06-4e2b-bc9a-049acf9ff730 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.250792] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.084s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.253714] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.598s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.255267] env[61728]: INFO nova.compute.claims [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.285991] env[61728]: INFO nova.scheduler.client.report [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Deleted allocations for instance 3ef9bf60-c6a1-4b7a-a375-5397fb871850 [ 876.388926] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Releasing lock "refresh_cache-121bef05-ef7b-47ee-b737-15f8b9d91158" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.389234] env[61728]: DEBUG nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Instance network_info: |[{"id": "58d83a36-5bb8-4f09-bd16-16d01ec692d7", "address": "fa:16:3e:59:8b:54", "network": {"id": "74344aaa-7298-4ae0-ab9c-ebeef230b69d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1018009435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9523aa656ff2459b94bb28d133fa4b05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58d83a36-5b", "ovs_interfaceid": "58d83a36-5bb8-4f09-bd16-16d01ec692d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 876.389720] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:8b:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9bb629cd-6d0f-4bed-965c-bd04a2f3ec49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '58d83a36-5bb8-4f09-bd16-16d01ec692d7', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.398522] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Creating folder: Project (9523aa656ff2459b94bb28d133fa4b05). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 876.399092] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e1fc4788-31a1-4826-af70-2e056243c587 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.412282] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Created folder: Project (9523aa656ff2459b94bb28d133fa4b05) in parent group-v121913. [ 876.412502] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Creating folder: Instances. Parent ref: group-v122131. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 876.412852] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52ff6910-dcc1-444d-a04a-be236d5e8a74 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.424250] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Created folder: Instances in parent group-v122131. [ 876.424749] env[61728]: DEBUG oslo.service.loopingcall [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.425017] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 876.425313] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2a2cce6f-2ef2-4169-9e12-269e3783b08c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.450038] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.450038] env[61728]: value = "task-464376" [ 876.450038] env[61728]: _type = "Task" [ 876.450038] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.459413] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464376, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.572727] env[61728]: DEBUG nova.compute.manager [req-cbc0b58c-86fb-43dd-a81e-7b5a68693dc1 req-ae288cf0-d091-4104-8ee5-650956f22eca service nova] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Received event network-vif-plugged-d16b0ff3-2873-4aa9-b5da-c97c6e5d0633 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 876.573110] env[61728]: DEBUG oslo_concurrency.lockutils [req-cbc0b58c-86fb-43dd-a81e-7b5a68693dc1 req-ae288cf0-d091-4104-8ee5-650956f22eca service nova] Acquiring lock "40558dd0-69d3-42d6-87cf-8e4ae4ae9dae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.573441] env[61728]: DEBUG oslo_concurrency.lockutils [req-cbc0b58c-86fb-43dd-a81e-7b5a68693dc1 req-ae288cf0-d091-4104-8ee5-650956f22eca service nova] Lock "40558dd0-69d3-42d6-87cf-8e4ae4ae9dae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.573750] env[61728]: DEBUG oslo_concurrency.lockutils [req-cbc0b58c-86fb-43dd-a81e-7b5a68693dc1 req-ae288cf0-d091-4104-8ee5-650956f22eca service nova] Lock "40558dd0-69d3-42d6-87cf-8e4ae4ae9dae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.574134] env[61728]: DEBUG nova.compute.manager [req-cbc0b58c-86fb-43dd-a81e-7b5a68693dc1 req-ae288cf0-d091-4104-8ee5-650956f22eca service nova] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] No waiting events found dispatching network-vif-plugged-d16b0ff3-2873-4aa9-b5da-c97c6e5d0633 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 876.574665] env[61728]: WARNING nova.compute.manager [req-cbc0b58c-86fb-43dd-a81e-7b5a68693dc1 req-ae288cf0-d091-4104-8ee5-650956f22eca service nova] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Received unexpected event network-vif-plugged-d16b0ff3-2873-4aa9-b5da-c97c6e5d0633 for instance with vm_state building and task_state spawning. [ 876.730024] env[61728]: DEBUG nova.network.neutron [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Successfully updated port: d16b0ff3-2873-4aa9-b5da-c97c6e5d0633 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 876.795029] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5cc074c9-ffd5-490c-9bb8-76c307c9f2a0 tempest-ImagesOneServerNegativeTestJSON-231395013 tempest-ImagesOneServerNegativeTestJSON-231395013-project-member] Lock "3ef9bf60-c6a1-4b7a-a375-5397fb871850" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.355s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.903653] env[61728]: DEBUG nova.compute.manager [req-1cdff35f-7fcb-4b00-98c2-00a4cef15d16 req-25ef791c-ba1e-4bda-85a1-c5003c0b20c4 service nova] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Received event network-changed-58d83a36-5bb8-4f09-bd16-16d01ec692d7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 876.904901] env[61728]: DEBUG nova.compute.manager [req-1cdff35f-7fcb-4b00-98c2-00a4cef15d16 req-25ef791c-ba1e-4bda-85a1-c5003c0b20c4 service nova] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Refreshing instance network info cache due to event network-changed-58d83a36-5bb8-4f09-bd16-16d01ec692d7. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 876.905173] env[61728]: DEBUG oslo_concurrency.lockutils [req-1cdff35f-7fcb-4b00-98c2-00a4cef15d16 req-25ef791c-ba1e-4bda-85a1-c5003c0b20c4 service nova] Acquiring lock "refresh_cache-121bef05-ef7b-47ee-b737-15f8b9d91158" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.905335] env[61728]: DEBUG oslo_concurrency.lockutils [req-1cdff35f-7fcb-4b00-98c2-00a4cef15d16 req-25ef791c-ba1e-4bda-85a1-c5003c0b20c4 service nova] Acquired lock "refresh_cache-121bef05-ef7b-47ee-b737-15f8b9d91158" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.905533] env[61728]: DEBUG nova.network.neutron [req-1cdff35f-7fcb-4b00-98c2-00a4cef15d16 req-25ef791c-ba1e-4bda-85a1-c5003c0b20c4 service nova] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Refreshing network info cache for port 58d83a36-5bb8-4f09-bd16-16d01ec692d7 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 876.961988] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464376, 'name': CreateVM_Task, 'duration_secs': 0.401607} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.962409] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 876.963702] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.964014] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.964549] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 876.964879] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6855a0d3-2888-4833-ad48-b253e0f0dd5d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.972322] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 876.972322] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52accb95-89aa-38f4-68fc-0f98b88e23b6" [ 876.972322] env[61728]: _type = "Task" [ 876.972322] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.982428] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52accb95-89aa-38f4-68fc-0f98b88e23b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.234586] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "refresh_cache-40558dd0-69d3-42d6-87cf-8e4ae4ae9dae" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.234957] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired lock "refresh_cache-40558dd0-69d3-42d6-87cf-8e4ae4ae9dae" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.235258] env[61728]: DEBUG nova.network.neutron [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 877.494097] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52accb95-89aa-38f4-68fc-0f98b88e23b6, 'name': SearchDatastore_Task, 'duration_secs': 0.012172} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.495144] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.495144] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.495551] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.495964] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.496587] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.500153] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-00d39cca-8588-4fc8-ba5c-21d0dc7114f8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.510576] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.511866] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 877.512626] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5df99dd4-8da3-430d-b0a2-6f73ddfaa68d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.525822] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 877.525822] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f40555-ee0d-15e1-d441-9d0e801094f2" [ 877.525822] env[61728]: _type = "Task" [ 877.525822] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.540171] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f40555-ee0d-15e1-d441-9d0e801094f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.697263] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a129ac-79b7-449f-8c13-ea9a092c71be {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.707325] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc749ae2-931b-4c9f-8906-cb1fb5c2225d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.748529] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84835058-a3da-4408-be35-1a571d61f422 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.758095] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe9643da-4361-4df3-90fb-53d056d9b59a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.775664] env[61728]: DEBUG nova.compute.provider_tree [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.794168] env[61728]: DEBUG nova.network.neutron [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 877.824378] env[61728]: DEBUG nova.network.neutron [req-1cdff35f-7fcb-4b00-98c2-00a4cef15d16 req-25ef791c-ba1e-4bda-85a1-c5003c0b20c4 service nova] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Updated VIF entry in instance network info cache for port 58d83a36-5bb8-4f09-bd16-16d01ec692d7. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 877.824739] env[61728]: DEBUG nova.network.neutron [req-1cdff35f-7fcb-4b00-98c2-00a4cef15d16 req-25ef791c-ba1e-4bda-85a1-c5003c0b20c4 service nova] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Updating instance_info_cache with network_info: [{"id": "58d83a36-5bb8-4f09-bd16-16d01ec692d7", "address": "fa:16:3e:59:8b:54", "network": {"id": "74344aaa-7298-4ae0-ab9c-ebeef230b69d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1018009435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9523aa656ff2459b94bb28d133fa4b05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58d83a36-5b", "ovs_interfaceid": "58d83a36-5bb8-4f09-bd16-16d01ec692d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.038531] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f40555-ee0d-15e1-d441-9d0e801094f2, 'name': SearchDatastore_Task, 'duration_secs': 0.015962} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.039891] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9edad8eb-9632-4c8a-905c-edfa119ff24a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.049027] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 878.049027] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5246898f-e5e9-97f1-0d4f-6f78a826b463" [ 878.049027] env[61728]: _type = "Task" [ 878.049027] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.057600] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5246898f-e5e9-97f1-0d4f-6f78a826b463, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.237437] env[61728]: DEBUG nova.network.neutron [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Updating instance_info_cache with network_info: [{"id": "d16b0ff3-2873-4aa9-b5da-c97c6e5d0633", "address": "fa:16:3e:c6:29:51", "network": {"id": "74344aaa-7298-4ae0-ab9c-ebeef230b69d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1018009435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9523aa656ff2459b94bb28d133fa4b05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd16b0ff3-28", "ovs_interfaceid": "d16b0ff3-2873-4aa9-b5da-c97c6e5d0633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.287221] env[61728]: DEBUG nova.scheduler.client.report [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 878.328441] env[61728]: DEBUG oslo_concurrency.lockutils [req-1cdff35f-7fcb-4b00-98c2-00a4cef15d16 req-25ef791c-ba1e-4bda-85a1-c5003c0b20c4 service nova] Releasing lock "refresh_cache-121bef05-ef7b-47ee-b737-15f8b9d91158" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.562497] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5246898f-e5e9-97f1-0d4f-6f78a826b463, 'name': SearchDatastore_Task, 'duration_secs': 0.017002} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.562983] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.563976] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 121bef05-ef7b-47ee-b737-15f8b9d91158/121bef05-ef7b-47ee-b737-15f8b9d91158.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 878.563976] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49dd00e6-ff9d-4918-8abf-90ede8dd69d9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.572534] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 878.572534] env[61728]: value = "task-464377" [ 878.572534] env[61728]: _type = "Task" [ 878.572534] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.582588] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464377, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.606405] env[61728]: DEBUG nova.compute.manager [req-aad30c31-6902-450e-adea-fa6188a54a1a req-2173f329-26df-4df7-aa84-9eab14c6cfca service nova] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Received event network-changed-d16b0ff3-2873-4aa9-b5da-c97c6e5d0633 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 878.606567] env[61728]: DEBUG nova.compute.manager [req-aad30c31-6902-450e-adea-fa6188a54a1a req-2173f329-26df-4df7-aa84-9eab14c6cfca service nova] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Refreshing instance network info cache due to event network-changed-d16b0ff3-2873-4aa9-b5da-c97c6e5d0633. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 878.606768] env[61728]: DEBUG oslo_concurrency.lockutils [req-aad30c31-6902-450e-adea-fa6188a54a1a req-2173f329-26df-4df7-aa84-9eab14c6cfca service nova] Acquiring lock "refresh_cache-40558dd0-69d3-42d6-87cf-8e4ae4ae9dae" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.741311] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Releasing lock "refresh_cache-40558dd0-69d3-42d6-87cf-8e4ae4ae9dae" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.744141] env[61728]: DEBUG nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Instance network_info: |[{"id": "d16b0ff3-2873-4aa9-b5da-c97c6e5d0633", "address": "fa:16:3e:c6:29:51", "network": {"id": "74344aaa-7298-4ae0-ab9c-ebeef230b69d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1018009435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9523aa656ff2459b94bb28d133fa4b05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd16b0ff3-28", "ovs_interfaceid": "d16b0ff3-2873-4aa9-b5da-c97c6e5d0633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.744141] env[61728]: DEBUG oslo_concurrency.lockutils [req-aad30c31-6902-450e-adea-fa6188a54a1a req-2173f329-26df-4df7-aa84-9eab14c6cfca service nova] Acquired lock "refresh_cache-40558dd0-69d3-42d6-87cf-8e4ae4ae9dae" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.744141] env[61728]: DEBUG nova.network.neutron [req-aad30c31-6902-450e-adea-fa6188a54a1a req-2173f329-26df-4df7-aa84-9eab14c6cfca service nova] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Refreshing network info cache for port d16b0ff3-2873-4aa9-b5da-c97c6e5d0633 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 878.744900] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:29:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9bb629cd-6d0f-4bed-965c-bd04a2f3ec49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd16b0ff3-2873-4aa9-b5da-c97c6e5d0633', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.758421] env[61728]: DEBUG oslo.service.loopingcall [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.762117] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 878.762117] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4c4fc28d-3d05-4d21-b37b-cc7f3035c17a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.793034] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.793034] env[61728]: DEBUG nova.compute.manager [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 878.795149] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.795149] env[61728]: value = "task-464378" [ 878.795149] env[61728]: _type = "Task" [ 878.795149] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.796023] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.377s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.797761] env[61728]: INFO nova.compute.claims [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 878.812811] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464378, 'name': CreateVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.088480] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464377, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.302562] env[61728]: DEBUG nova.compute.utils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.308883] env[61728]: DEBUG nova.compute.manager [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 879.309693] env[61728]: DEBUG nova.network.neutron [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 879.332738] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464378, 'name': CreateVM_Task, 'duration_secs': 0.437275} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.337548] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 879.338361] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.338540] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.338869] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 879.339441] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a9a94f3-d474-404e-90b8-be55aa1a0230 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.346018] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 879.346018] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f63a01-96d5-c728-1731-2ae319e7dfab" [ 879.346018] env[61728]: _type = "Task" [ 879.346018] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.357560] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f63a01-96d5-c728-1731-2ae319e7dfab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.436287] env[61728]: DEBUG nova.policy [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e35028155c743e4bcf2ab91be37b201', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c035f75f40864559b47b643fcfca56e1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 879.590648] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464377, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.685899} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.593630] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 121bef05-ef7b-47ee-b737-15f8b9d91158/121bef05-ef7b-47ee-b737-15f8b9d91158.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 879.593969] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.594623] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1fd1add3-eac6-420f-9249-8bcce1fc6505 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.606751] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 879.606751] env[61728]: value = "task-464379" [ 879.606751] env[61728]: _type = "Task" [ 879.606751] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.620949] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464379, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.649218] env[61728]: DEBUG nova.network.neutron [req-aad30c31-6902-450e-adea-fa6188a54a1a req-2173f329-26df-4df7-aa84-9eab14c6cfca service nova] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Updated VIF entry in instance network info cache for port d16b0ff3-2873-4aa9-b5da-c97c6e5d0633. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 879.649601] env[61728]: DEBUG nova.network.neutron [req-aad30c31-6902-450e-adea-fa6188a54a1a req-2173f329-26df-4df7-aa84-9eab14c6cfca service nova] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Updating instance_info_cache with network_info: [{"id": "d16b0ff3-2873-4aa9-b5da-c97c6e5d0633", "address": "fa:16:3e:c6:29:51", "network": {"id": "74344aaa-7298-4ae0-ab9c-ebeef230b69d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1018009435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9523aa656ff2459b94bb28d133fa4b05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd16b0ff3-28", "ovs_interfaceid": "d16b0ff3-2873-4aa9-b5da-c97c6e5d0633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.819871] env[61728]: DEBUG nova.compute.manager [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 879.859532] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f63a01-96d5-c728-1731-2ae319e7dfab, 'name': SearchDatastore_Task, 'duration_secs': 0.015087} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.859750] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.860015] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.860268] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.860419] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.861473] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.861473] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dce64e0d-ced5-4243-a244-99c9f89381c7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.872570] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.872831] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 879.873646] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46706297-5198-45ad-a85c-25cd8e41698f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.882505] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 879.882505] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52417f6d-dd40-ec4a-d039-58ba11fcfb06" [ 879.882505] env[61728]: _type = "Task" [ 879.882505] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.890144] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52417f6d-dd40-ec4a-d039-58ba11fcfb06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.049235] env[61728]: DEBUG nova.network.neutron [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Successfully created port: 4a6313df-8741-4533-ae0e-1f469193c389 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 880.128800] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464379, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078936} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.129097] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 880.132621] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8188b3c4-0fca-4dbf-a36f-83dc5b8d8412 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.163658] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 121bef05-ef7b-47ee-b737-15f8b9d91158/121bef05-ef7b-47ee-b737-15f8b9d91158.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.167465] env[61728]: DEBUG oslo_concurrency.lockutils [req-aad30c31-6902-450e-adea-fa6188a54a1a req-2173f329-26df-4df7-aa84-9eab14c6cfca service nova] Releasing lock "refresh_cache-40558dd0-69d3-42d6-87cf-8e4ae4ae9dae" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.168353] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48115e1d-4f9a-487b-bc5f-d4d15e63eef6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.191552] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 880.191552] env[61728]: value = "task-464380" [ 880.191552] env[61728]: _type = "Task" [ 880.191552] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.211393] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464380, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.370759] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c83cec8-1eb2-462a-a942-1a4d8219ab76 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.379775] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830129a5-9a9b-4738-9943-b35fb6a8bc08 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.396642] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52417f6d-dd40-ec4a-d039-58ba11fcfb06, 'name': SearchDatastore_Task, 'duration_secs': 0.016266} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.421647] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de60879d-b565-461e-bdea-e4a177f13e1e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.425210] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfca434c-7bc9-4b98-8132-f0b236fbdd48 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.434117] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 880.434117] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5290e908-1e21-c97a-b191-e82700905e44" [ 880.434117] env[61728]: _type = "Task" [ 880.434117] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.435517] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c461120-60e4-4f03-b08d-35b5aca946ea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.462295] env[61728]: DEBUG nova.compute.provider_tree [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 880.463759] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5290e908-1e21-c97a-b191-e82700905e44, 'name': SearchDatastore_Task, 'duration_secs': 0.014118} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.464432] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.464432] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae/40558dd0-69d3-42d6-87cf-8e4ae4ae9dae.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 880.465143] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fe65084a-75b5-40d0-9c15-5bd87810b00b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.473038] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 880.473038] env[61728]: value = "task-464381" [ 880.473038] env[61728]: _type = "Task" [ 880.473038] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.483100] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464381, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.704455] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464380, 'name': ReconfigVM_Task, 'duration_secs': 0.301306} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.704789] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 121bef05-ef7b-47ee-b737-15f8b9d91158/121bef05-ef7b-47ee-b737-15f8b9d91158.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.706260] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5abae49a-170b-4359-b1bf-94c8b3659375 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.713856] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 880.713856] env[61728]: value = "task-464382" [ 880.713856] env[61728]: _type = "Task" [ 880.713856] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.722061] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464382, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.846999] env[61728]: DEBUG nova.compute.manager [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 880.877063] env[61728]: DEBUG nova.virt.hardware [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.877355] env[61728]: DEBUG nova.virt.hardware [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.877523] env[61728]: DEBUG nova.virt.hardware [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.877714] env[61728]: DEBUG nova.virt.hardware [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.877896] env[61728]: DEBUG nova.virt.hardware [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.878095] env[61728]: DEBUG nova.virt.hardware [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.878335] env[61728]: DEBUG nova.virt.hardware [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.878517] env[61728]: DEBUG nova.virt.hardware [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.878680] env[61728]: DEBUG nova.virt.hardware [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.878872] env[61728]: DEBUG nova.virt.hardware [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.879102] env[61728]: DEBUG nova.virt.hardware [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.880097] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc4ab63-00e2-4b65-b9a2-9206c0dd9936 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.888536] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d43e16af-2aaf-45ec-beb5-80d97e9271de {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.982813] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464381, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.988079] env[61728]: ERROR nova.scheduler.client.report [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [req-a3a292f7-9939-4305-93bb-76e31bc657c2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a3a292f7-9939-4305-93bb-76e31bc657c2"}]} [ 881.004374] env[61728]: DEBUG nova.scheduler.client.report [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 881.020632] env[61728]: DEBUG nova.scheduler.client.report [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 881.020949] env[61728]: DEBUG nova.compute.provider_tree [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 881.034212] env[61728]: DEBUG nova.scheduler.client.report [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 881.055040] env[61728]: DEBUG nova.scheduler.client.report [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 881.232737] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464382, 'name': Rename_Task, 'duration_secs': 0.233293} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.232737] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 881.235850] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ea7a243-28a7-4109-b40a-f7bd4f7e98aa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.243607] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 881.243607] env[61728]: value = "task-464383" [ 881.243607] env[61728]: _type = "Task" [ 881.243607] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.258062] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464383, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.486701] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464381, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.855513} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.488380] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae/40558dd0-69d3-42d6-87cf-8e4ae4ae9dae.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 881.488688] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.489569] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196b6366-1c96-434d-b1bd-c64072ec569f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.492901] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d4ee07ec-8629-40c5-8dd2-46de20d82ee6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.500686] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf2d6b4-289a-40b0-a386-2e20e631ba86 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.506765] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 881.506765] env[61728]: value = "task-464384" [ 881.506765] env[61728]: _type = "Task" [ 881.506765] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.542032] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7dab858-2326-4ea3-9d51-725a5e322241 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.548291] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464384, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.553594] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a248976-2ddf-49f7-b221-af2e9869e60c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.569724] env[61728]: DEBUG nova.compute.provider_tree [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 881.757417] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464383, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.956684] env[61728]: DEBUG nova.compute.manager [req-61bef8e4-0f6f-4750-95ad-9fd9f7d0d996 req-e2ae2a70-a984-44fb-b92c-dde8d30643c3 service nova] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Received event network-vif-plugged-4a6313df-8741-4533-ae0e-1f469193c389 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 881.956964] env[61728]: DEBUG oslo_concurrency.lockutils [req-61bef8e4-0f6f-4750-95ad-9fd9f7d0d996 req-e2ae2a70-a984-44fb-b92c-dde8d30643c3 service nova] Acquiring lock "f27f22aa-f05d-4981-a389-311731c51f93-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.957241] env[61728]: DEBUG oslo_concurrency.lockutils [req-61bef8e4-0f6f-4750-95ad-9fd9f7d0d996 req-e2ae2a70-a984-44fb-b92c-dde8d30643c3 service nova] Lock "f27f22aa-f05d-4981-a389-311731c51f93-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.957461] env[61728]: DEBUG oslo_concurrency.lockutils [req-61bef8e4-0f6f-4750-95ad-9fd9f7d0d996 req-e2ae2a70-a984-44fb-b92c-dde8d30643c3 service nova] Lock "f27f22aa-f05d-4981-a389-311731c51f93-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.957669] env[61728]: DEBUG nova.compute.manager [req-61bef8e4-0f6f-4750-95ad-9fd9f7d0d996 req-e2ae2a70-a984-44fb-b92c-dde8d30643c3 service nova] [instance: f27f22aa-f05d-4981-a389-311731c51f93] No waiting events found dispatching network-vif-plugged-4a6313df-8741-4533-ae0e-1f469193c389 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 881.957923] env[61728]: WARNING nova.compute.manager [req-61bef8e4-0f6f-4750-95ad-9fd9f7d0d996 req-e2ae2a70-a984-44fb-b92c-dde8d30643c3 service nova] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Received unexpected event network-vif-plugged-4a6313df-8741-4533-ae0e-1f469193c389 for instance with vm_state building and task_state spawning. [ 882.017198] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464384, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073956} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.017502] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.019627] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb2cdad-c293-4524-b38b-b0a3d3afb065 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.023486] env[61728]: DEBUG nova.network.neutron [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Successfully updated port: 4a6313df-8741-4533-ae0e-1f469193c389 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 882.044593] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae/40558dd0-69d3-42d6-87cf-8e4ae4ae9dae.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.046006] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.046157] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.046306] env[61728]: DEBUG nova.network.neutron [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 882.047381] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-275abb7e-09f3-4aee-a587-1bc92eefda3d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.070628] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 882.070628] env[61728]: value = "task-464385" [ 882.070628] env[61728]: _type = "Task" [ 882.070628] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.081993] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464385, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.090808] env[61728]: ERROR nova.scheduler.client.report [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [req-fd294ede-c5e6-4863-ba88-5f0a3fee8d92] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-fd294ede-c5e6-4863-ba88-5f0a3fee8d92"}]} [ 882.093724] env[61728]: DEBUG nova.network.neutron [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 882.113187] env[61728]: DEBUG nova.scheduler.client.report [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 882.128152] env[61728]: DEBUG nova.scheduler.client.report [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 882.128386] env[61728]: DEBUG nova.compute.provider_tree [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 882.143562] env[61728]: DEBUG nova.scheduler.client.report [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 882.164797] env[61728]: DEBUG nova.scheduler.client.report [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 882.256341] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464383, 'name': PowerOnVM_Task, 'duration_secs': 0.516505} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.256341] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 882.256341] env[61728]: INFO nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Took 8.60 seconds to spawn the instance on the hypervisor. [ 882.256680] env[61728]: DEBUG nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.261132] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1164ef-fb5b-444e-84b5-6835582351bb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.310734] env[61728]: DEBUG nova.network.neutron [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Updating instance_info_cache with network_info: [{"id": "4a6313df-8741-4533-ae0e-1f469193c389", "address": "fa:16:3e:43:88:0c", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a6313df-87", "ovs_interfaceid": "4a6313df-8741-4533-ae0e-1f469193c389", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.310995] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.311248] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.552231] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdca8a94-28ab-433a-94d1-2e0a5f7d6d84 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.560115] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e780c272-b9c5-4395-a9de-fa535b9f8dff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.594056] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7f7dc6-9a08-4133-8dba-54a4154d16e2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.601966] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464385, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.607937] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f0076e9-fb29-443c-99d1-798319460e0e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.624457] env[61728]: DEBUG nova.compute.provider_tree [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 882.786098] env[61728]: INFO nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Took 39.00 seconds to build instance. [ 882.812919] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.812919] env[61728]: DEBUG nova.compute.manager [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Instance network_info: |[{"id": "4a6313df-8741-4533-ae0e-1f469193c389", "address": "fa:16:3e:43:88:0c", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a6313df-87", "ovs_interfaceid": "4a6313df-8741-4533-ae0e-1f469193c389", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 882.814034] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:88:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2be3fdb5-359e-43bd-8c20-2ff00e81db55', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4a6313df-8741-4533-ae0e-1f469193c389', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 882.822499] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Creating folder: Project (c035f75f40864559b47b643fcfca56e1). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 882.823229] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d95f7129-bf98-4b86-9542-6d48bfae0283 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.834249] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Created folder: Project (c035f75f40864559b47b643fcfca56e1) in parent group-v121913. [ 882.835296] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Creating folder: Instances. Parent ref: group-v122135. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 882.835296] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d47d8526-82dc-473d-8c15-bcd9ea1ad789 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.844399] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Created folder: Instances in parent group-v122135. [ 882.844399] env[61728]: DEBUG oslo.service.loopingcall [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.844399] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 882.844581] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6acc918e-c1bd-4f1e-8056-54bfb88e83bf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.864390] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 882.864390] env[61728]: value = "task-464388" [ 882.864390] env[61728]: _type = "Task" [ 882.864390] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.874203] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464388, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.099923] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464385, 'name': ReconfigVM_Task, 'duration_secs': 0.645892} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.100251] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae/40558dd0-69d3-42d6-87cf-8e4ae4ae9dae.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.100888] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4f008942-c25f-4a7d-a3e7-fce2736c0277 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.107869] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 883.107869] env[61728]: value = "task-464389" [ 883.107869] env[61728]: _type = "Task" [ 883.107869] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.115814] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464389, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.161208] env[61728]: DEBUG nova.scheduler.client.report [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 127 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 883.161643] env[61728]: DEBUG nova.compute.provider_tree [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 127 to 128 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 883.161934] env[61728]: DEBUG nova.compute.provider_tree [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 883.287374] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "121bef05-ef7b-47ee-b737-15f8b9d91158" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.515s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.374945] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464388, 'name': CreateVM_Task, 'duration_secs': 0.380126} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.375171] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 883.375837] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.376019] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.376370] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 883.376636] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b195da3e-506a-4819-a86c-c8d5bbe2e507 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.381541] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 883.381541] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524572ee-f4f4-f54b-5742-f36c093832aa" [ 883.381541] env[61728]: _type = "Task" [ 883.381541] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.389298] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524572ee-f4f4-f54b-5742-f36c093832aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.620802] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464389, 'name': Rename_Task, 'duration_secs': 0.206347} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.620802] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 883.621105] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df5c7c6d-75fe-45b3-8d95-9dc7988c9190 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.628963] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 883.628963] env[61728]: value = "task-464390" [ 883.628963] env[61728]: _type = "Task" [ 883.628963] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.638619] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464390, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.669030] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.873s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.669569] env[61728]: DEBUG nova.compute.manager [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 883.672860] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.066s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.675431] env[61728]: INFO nova.compute.claims [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 883.790316] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 883.893098] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524572ee-f4f4-f54b-5742-f36c093832aa, 'name': SearchDatastore_Task, 'duration_secs': 0.01431} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.893356] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.893580] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.893846] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.894019] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.894201] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.894491] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02fb4772-63ad-4859-b188-4675b98808c7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.903493] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.903661] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 883.904412] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88756570-2150-40f1-8b2a-662ce99a0600 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.911382] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 883.911382] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5287f06e-9667-431c-bdfe-82d876b26705" [ 883.911382] env[61728]: _type = "Task" [ 883.911382] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.919608] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5287f06e-9667-431c-bdfe-82d876b26705, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.991572] env[61728]: DEBUG nova.compute.manager [req-c6231f55-4ab3-49c4-b701-1ec9e1f63057 req-f232231a-a9f0-4b6b-9c59-ecbd0ebfe823 service nova] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Received event network-changed-4a6313df-8741-4533-ae0e-1f469193c389 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 883.991933] env[61728]: DEBUG nova.compute.manager [req-c6231f55-4ab3-49c4-b701-1ec9e1f63057 req-f232231a-a9f0-4b6b-9c59-ecbd0ebfe823 service nova] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Refreshing instance network info cache due to event network-changed-4a6313df-8741-4533-ae0e-1f469193c389. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 883.992295] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6231f55-4ab3-49c4-b701-1ec9e1f63057 req-f232231a-a9f0-4b6b-9c59-ecbd0ebfe823 service nova] Acquiring lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.992571] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6231f55-4ab3-49c4-b701-1ec9e1f63057 req-f232231a-a9f0-4b6b-9c59-ecbd0ebfe823 service nova] Acquired lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.992884] env[61728]: DEBUG nova.network.neutron [req-c6231f55-4ab3-49c4-b701-1ec9e1f63057 req-f232231a-a9f0-4b6b-9c59-ecbd0ebfe823 service nova] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Refreshing network info cache for port 4a6313df-8741-4533-ae0e-1f469193c389 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 884.138509] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464390, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.181396] env[61728]: DEBUG nova.compute.utils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 884.183121] env[61728]: DEBUG nova.compute.manager [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 884.183346] env[61728]: DEBUG nova.network.neutron [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 884.320802] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.320802] env[61728]: DEBUG nova.policy [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e35028155c743e4bcf2ab91be37b201', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c035f75f40864559b47b643fcfca56e1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 884.422530] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5287f06e-9667-431c-bdfe-82d876b26705, 'name': SearchDatastore_Task, 'duration_secs': 0.011057} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.423015] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbb2b7f8-7bd6-4353-875b-6eb6efbc5b77 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.428971] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 884.428971] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52624253-e5ee-4b09-6c98-6df3f0184d21" [ 884.428971] env[61728]: _type = "Task" [ 884.428971] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.437251] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52624253-e5ee-4b09-6c98-6df3f0184d21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.624508] env[61728]: DEBUG nova.network.neutron [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Successfully created port: 32361465-1ad4-4a38-9989-8110f57692f7 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 884.638944] env[61728]: DEBUG oslo_vmware.api [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464390, 'name': PowerOnVM_Task, 'duration_secs': 0.74156} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.639362] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 884.639677] env[61728]: INFO nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Took 8.44 seconds to spawn the instance on the hypervisor. [ 884.639915] env[61728]: DEBUG nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.640849] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7308a662-dc6d-4775-bcf3-420416d7e3f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.689018] env[61728]: DEBUG nova.compute.manager [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 884.708020] env[61728]: DEBUG nova.network.neutron [req-c6231f55-4ab3-49c4-b701-1ec9e1f63057 req-f232231a-a9f0-4b6b-9c59-ecbd0ebfe823 service nova] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Updated VIF entry in instance network info cache for port 4a6313df-8741-4533-ae0e-1f469193c389. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 884.708441] env[61728]: DEBUG nova.network.neutron [req-c6231f55-4ab3-49c4-b701-1ec9e1f63057 req-f232231a-a9f0-4b6b-9c59-ecbd0ebfe823 service nova] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Updating instance_info_cache with network_info: [{"id": "4a6313df-8741-4533-ae0e-1f469193c389", "address": "fa:16:3e:43:88:0c", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a6313df-87", "ovs_interfaceid": "4a6313df-8741-4533-ae0e-1f469193c389", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.811196] env[61728]: DEBUG oslo_vmware.rw_handles [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5284986e-eff9-c129-3129-76bc62dac282/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 884.812151] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c40312-f53b-489b-b661-ff6154b78e2e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.822185] env[61728]: DEBUG oslo_vmware.rw_handles [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5284986e-eff9-c129-3129-76bc62dac282/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 884.822403] env[61728]: ERROR oslo_vmware.rw_handles [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5284986e-eff9-c129-3129-76bc62dac282/disk-0.vmdk due to incomplete transfer. [ 884.822644] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6911af8c-2221-4ebb-bc29-e7d37bb950d3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.829435] env[61728]: DEBUG oslo_vmware.rw_handles [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5284986e-eff9-c129-3129-76bc62dac282/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 884.829679] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Uploaded image 6bf918f6-e123-44bd-a072-fe70029d1786 to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 884.831998] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 884.832325] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a9df7645-8210-4667-988d-525c70aa8289 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.838088] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 884.838088] env[61728]: value = "task-464391" [ 884.838088] env[61728]: _type = "Task" [ 884.838088] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.851277] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464391, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.942411] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52624253-e5ee-4b09-6c98-6df3f0184d21, 'name': SearchDatastore_Task, 'duration_secs': 0.011867} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.946227] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.946655] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] f27f22aa-f05d-4981-a389-311731c51f93/f27f22aa-f05d-4981-a389-311731c51f93.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 884.947169] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6e64b872-6f77-4b09-bf87-d78b7e7a267a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.954627] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 884.954627] env[61728]: value = "task-464392" [ 884.954627] env[61728]: _type = "Task" [ 884.954627] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.969564] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464392, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.107036] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646fa1a0-62ce-41f8-9c39-dcf664251987 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.115335] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd6717c-2a31-4eab-af4f-b8167c25edd6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.165444] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd4723d-b306-4909-a705-d11d7d2eeedf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.167214] env[61728]: INFO nova.compute.manager [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Took 41.35 seconds to build instance. [ 885.174707] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b3f1421-514a-497c-85b5-a5e7316cdacc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.191389] env[61728]: DEBUG nova.compute.provider_tree [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.212236] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6231f55-4ab3-49c4-b701-1ec9e1f63057 req-f232231a-a9f0-4b6b-9c59-ecbd0ebfe823 service nova] Releasing lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.349579] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464391, 'name': Destroy_Task, 'duration_secs': 0.355312} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.349579] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Destroyed the VM [ 885.350535] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 885.350973] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-efda8b40-1a62-480b-b7a6-704ddd6f76d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.361526] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 885.361526] env[61728]: value = "task-464393" [ 885.361526] env[61728]: _type = "Task" [ 885.361526] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.372967] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464393, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.465884] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464392, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.670641] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d009f83d-a453-4d8d-9a5a-33c5131e9333 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "40558dd0-69d3-42d6-87cf-8e4ae4ae9dae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.866s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.694540] env[61728]: DEBUG nova.scheduler.client.report [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 885.700253] env[61728]: DEBUG nova.compute.manager [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 885.731603] env[61728]: DEBUG nova.virt.hardware [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 885.731945] env[61728]: DEBUG nova.virt.hardware [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 885.732130] env[61728]: DEBUG nova.virt.hardware [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.732324] env[61728]: DEBUG nova.virt.hardware [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 885.732472] env[61728]: DEBUG nova.virt.hardware [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.732640] env[61728]: DEBUG nova.virt.hardware [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 885.732887] env[61728]: DEBUG nova.virt.hardware [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 885.733074] env[61728]: DEBUG nova.virt.hardware [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 885.733257] env[61728]: DEBUG nova.virt.hardware [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 885.733429] env[61728]: DEBUG nova.virt.hardware [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 885.733604] env[61728]: DEBUG nova.virt.hardware [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 885.734744] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-066a442c-0458-4221-b290-812fcc6b66d4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.743034] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05dbc356-006b-49d6-9a37-6ed0ba0914eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.871155] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464393, 'name': RemoveSnapshot_Task, 'duration_secs': 0.499938} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.872054] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 885.872054] env[61728]: DEBUG nova.compute.manager [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.872872] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa9b297e-a566-467d-a3e1-c0ab3a5d8775 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.966657] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464392, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548342} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.967268] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] f27f22aa-f05d-4981-a389-311731c51f93/f27f22aa-f05d-4981-a389-311731c51f93.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 885.968285] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 885.968594] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18f4c8bc-18a6-4f60-9543-6477d10154ee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.975529] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 885.975529] env[61728]: value = "task-464394" [ 885.975529] env[61728]: _type = "Task" [ 885.975529] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.985398] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464394, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.173390] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 886.200468] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.528s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.200999] env[61728]: DEBUG nova.compute.manager [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 886.204950] env[61728]: DEBUG oslo_concurrency.lockutils [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.908s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.205221] env[61728]: DEBUG oslo_concurrency.lockutils [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.207391] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.585s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.207621] env[61728]: DEBUG nova.objects.instance [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lazy-loading 'resources' on Instance uuid d44718bd-872b-401c-aa11-f10bea4a35d8 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.209296] env[61728]: DEBUG oslo_concurrency.lockutils [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "40558dd0-69d3-42d6-87cf-8e4ae4ae9dae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.209473] env[61728]: DEBUG oslo_concurrency.lockutils [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "40558dd0-69d3-42d6-87cf-8e4ae4ae9dae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.209684] env[61728]: DEBUG oslo_concurrency.lockutils [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "40558dd0-69d3-42d6-87cf-8e4ae4ae9dae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.209873] env[61728]: DEBUG oslo_concurrency.lockutils [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "40558dd0-69d3-42d6-87cf-8e4ae4ae9dae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.210053] env[61728]: DEBUG oslo_concurrency.lockutils [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "40558dd0-69d3-42d6-87cf-8e4ae4ae9dae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.213675] env[61728]: INFO nova.compute.manager [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Terminating instance [ 886.215978] env[61728]: DEBUG nova.compute.manager [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 886.216214] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 886.217416] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6256dd43-e00f-4fc7-b222-ca7c546a05c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.227284] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 886.227547] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2cc72cc8-7a7c-4e2b-8735-f03faf5200f6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.237158] env[61728]: DEBUG oslo_vmware.api [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 886.237158] env[61728]: value = "task-464395" [ 886.237158] env[61728]: _type = "Task" [ 886.237158] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.241170] env[61728]: INFO nova.scheduler.client.report [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleted allocations for instance f2692e22-6c14-4bb3-a080-607f0731105d [ 886.253447] env[61728]: DEBUG oslo_vmware.api [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464395, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.287397] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Acquiring lock "174607a1-9bc8-4e07-8993-7f0bb0f308e9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.287687] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Lock "174607a1-9bc8-4e07-8993-7f0bb0f308e9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.287877] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Acquiring lock "174607a1-9bc8-4e07-8993-7f0bb0f308e9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.288095] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Lock "174607a1-9bc8-4e07-8993-7f0bb0f308e9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.288287] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Lock "174607a1-9bc8-4e07-8993-7f0bb0f308e9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.290594] env[61728]: INFO nova.compute.manager [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Terminating instance [ 887.026584] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "121bef05-ef7b-47ee-b737-15f8b9d91158" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.026950] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "121bef05-ef7b-47ee-b737-15f8b9d91158" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.027031] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "121bef05-ef7b-47ee-b737-15f8b9d91158-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.027217] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "121bef05-ef7b-47ee-b737-15f8b9d91158-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.027387] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "121bef05-ef7b-47ee-b737-15f8b9d91158-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.034343] env[61728]: DEBUG nova.compute.utils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 887.040197] env[61728]: DEBUG nova.compute.manager [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.040197] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 887.040548] env[61728]: INFO nova.compute.manager [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Terminating instance [ 887.042932] env[61728]: DEBUG nova.network.neutron [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Successfully updated port: 32361465-1ad4-4a38-9989-8110f57692f7 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 887.043730] env[61728]: INFO nova.compute.manager [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Shelve offloading [ 887.050109] env[61728]: DEBUG nova.compute.manager [req-9984fd93-137e-419f-aaae-321d772f9746 req-fcaebb41-3691-45ad-9a34-a25a8fad532b service nova] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Received event network-vif-plugged-32361465-1ad4-4a38-9989-8110f57692f7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 887.050109] env[61728]: DEBUG oslo_concurrency.lockutils [req-9984fd93-137e-419f-aaae-321d772f9746 req-fcaebb41-3691-45ad-9a34-a25a8fad532b service nova] Acquiring lock "5fa2e9f4-6d68-46c2-a549-95d216bab886-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.050109] env[61728]: DEBUG oslo_concurrency.lockutils [req-9984fd93-137e-419f-aaae-321d772f9746 req-fcaebb41-3691-45ad-9a34-a25a8fad532b service nova] Lock "5fa2e9f4-6d68-46c2-a549-95d216bab886-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.050109] env[61728]: DEBUG oslo_concurrency.lockutils [req-9984fd93-137e-419f-aaae-321d772f9746 req-fcaebb41-3691-45ad-9a34-a25a8fad532b service nova] Lock "5fa2e9f4-6d68-46c2-a549-95d216bab886-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.050109] env[61728]: DEBUG nova.compute.manager [req-9984fd93-137e-419f-aaae-321d772f9746 req-fcaebb41-3691-45ad-9a34-a25a8fad532b service nova] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] No waiting events found dispatching network-vif-plugged-32361465-1ad4-4a38-9989-8110f57692f7 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 887.050109] env[61728]: WARNING nova.compute.manager [req-9984fd93-137e-419f-aaae-321d772f9746 req-fcaebb41-3691-45ad-9a34-a25a8fad532b service nova] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Received unexpected event network-vif-plugged-32361465-1ad4-4a38-9989-8110f57692f7 for instance with vm_state building and task_state spawning. [ 887.052487] env[61728]: DEBUG nova.compute.manager [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 887.052666] env[61728]: DEBUG nova.network.neutron [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 887.060497] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75dca4b5-3f4f-48e3-8c6e-e06db576d4df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.061079] env[61728]: DEBUG nova.compute.manager [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.061278] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 887.061897] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "refresh_cache-5fa2e9f4-6d68-46c2-a549-95d216bab886" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.062030] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "refresh_cache-5fa2e9f4-6d68-46c2-a549-95d216bab886" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.062167] env[61728]: DEBUG nova.network.neutron [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 887.063638] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 887.064576] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4045ed3d-4fb9-47eb-b26a-b42e60bbd9e2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.067448] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fbcbf7f9-4fad-4d7b-9f35-c0ab3cb7c2ef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.078140] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464394, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069296} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.085353] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 887.086123] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 887.086642] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 887.086642] env[61728]: value = "task-464396" [ 887.086642] env[61728]: _type = "Task" [ 887.086642] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.086861] env[61728]: DEBUG oslo_vmware.api [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464395, 'name': PowerOffVM_Task, 'duration_secs': 0.598896} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.087109] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 887.087810] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d6526f0-62ec-4656-91be-7b20637b57a5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.090645] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2b510718-11ba-4296-9bd5-5d81e578d432 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.092470] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 887.092655] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 887.092848] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2fdec026-8034-49b2-8153-7bc83a0b3fa7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.097110] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf7743b4-1a4e-400e-abd3-345fbf1ee3e3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.122802] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] f27f22aa-f05d-4981-a389-311731c51f93/f27f22aa-f05d-4981-a389-311731c51f93.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 887.131983] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1836c000-4e1a-4ef9-8061-dfebc46286ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.148014] env[61728]: DEBUG oslo_vmware.api [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Waiting for the task: (returnval){ [ 887.148014] env[61728]: value = "task-464398" [ 887.148014] env[61728]: _type = "Task" [ 887.148014] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.148336] env[61728]: DEBUG oslo_vmware.api [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 887.148336] env[61728]: value = "task-464397" [ 887.148336] env[61728]: _type = "Task" [ 887.148336] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.150834] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 887.150834] env[61728]: DEBUG nova.compute.manager [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.150834] env[61728]: DEBUG oslo_concurrency.lockutils [None req-346089e8-ebf3-4d28-a429-1f0c405d6ca4 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "f2692e22-6c14-4bb3-a080-607f0731105d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.087s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.154770] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f382cd5-bfee-4d01-ae76-4958e7605941 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.171535] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 887.171535] env[61728]: value = "task-464400" [ 887.171535] env[61728]: _type = "Task" [ 887.171535] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.174368] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 887.174619] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 887.174793] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Deleting the datastore file [datastore1] 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.178229] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ddaae353-c472-46be-88bf-8375651fe161 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.188301] env[61728]: DEBUG oslo_concurrency.lockutils [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.188495] env[61728]: DEBUG oslo_concurrency.lockutils [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquired lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.188684] env[61728]: DEBUG nova.network.neutron [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 887.189988] env[61728]: DEBUG oslo_vmware.api [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464398, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.192072] env[61728]: DEBUG oslo_vmware.api [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464397, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.192072] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.192788] env[61728]: DEBUG nova.policy [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '38f64fd31712472a841f82c620d9340b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f5329f59d8bf43b2a855c4259387fccd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 887.201933] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464400, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.203346] env[61728]: DEBUG oslo_vmware.api [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 887.203346] env[61728]: value = "task-464401" [ 887.203346] env[61728]: _type = "Task" [ 887.203346] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.212829] env[61728]: DEBUG oslo_vmware.api [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464401, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.513112] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afecc5ac-0b7d-4f89-8704-84e2dbcfcb5f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.529255] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839aa5a5-7a02-4b4c-a5ee-29c964297a6a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.559638] env[61728]: DEBUG nova.compute.manager [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 887.562900] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d41e1c-4653-4a48-9340-fece85b99eb5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.570547] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e8fb8a-fc64-4377-8141-849b547ef9d9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.589301] env[61728]: DEBUG nova.compute.provider_tree [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.603281] env[61728]: DEBUG nova.network.neutron [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 887.667504] env[61728]: DEBUG oslo_vmware.api [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464397, 'name': PowerOffVM_Task, 'duration_secs': 0.205848} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.667504] env[61728]: DEBUG oslo_vmware.api [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464398, 'name': PowerOffVM_Task, 'duration_secs': 0.202176} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.667662] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 887.667746] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 887.668020] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 887.668179] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 887.668406] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd7b0671-5190-44a8-b2c0-de162211d1b9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.670147] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ba578744-13c4-41e3-96f4-403da12c602e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.685141] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464400, 'name': ReconfigVM_Task, 'duration_secs': 0.305793} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.685431] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Reconfigured VM instance instance-0000004d to attach disk [datastore1] f27f22aa-f05d-4981-a389-311731c51f93/f27f22aa-f05d-4981-a389-311731c51f93.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.686078] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-86fd1214-2d93-40fd-b657-9b984485f43c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.694164] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 887.694164] env[61728]: value = "task-464404" [ 887.694164] env[61728]: _type = "Task" [ 887.694164] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.703711] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464404, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.704544] env[61728]: DEBUG nova.network.neutron [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Successfully created port: fbf3382b-a343-4cc4-9233-ebf4ac3139a4 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 887.715063] env[61728]: DEBUG oslo_vmware.api [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464401, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162693} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.715315] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 887.715501] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 887.715690] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 887.715866] env[61728]: INFO nova.compute.manager [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Took 1.50 seconds to destroy the instance on the hypervisor. [ 887.716147] env[61728]: DEBUG oslo.service.loopingcall [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.716381] env[61728]: DEBUG nova.compute.manager [-] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 887.716508] env[61728]: DEBUG nova.network.neutron [-] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 887.856404] env[61728]: DEBUG nova.network.neutron [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Updating instance_info_cache with network_info: [{"id": "32361465-1ad4-4a38-9989-8110f57692f7", "address": "fa:16:3e:c3:49:72", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32361465-1a", "ovs_interfaceid": "32361465-1ad4-4a38-9989-8110f57692f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.056912] env[61728]: DEBUG nova.network.neutron [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Updating instance_info_cache with network_info: [{"id": "124d4780-15ec-4f17-a2c4-edf1517ed589", "address": "fa:16:3e:6d:47:93", "network": {"id": "6a1908cc-f6aa-4dd7-970e-c8a07e1b16ee", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2047209328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "01839050e56e4dbfa6e22471a3c60998", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap124d4780-15", "ovs_interfaceid": "124d4780-15ec-4f17-a2c4-edf1517ed589", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.092455] env[61728]: DEBUG nova.scheduler.client.report [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.140978] env[61728]: DEBUG nova.compute.manager [req-797cc5aa-d65e-4535-89a3-590bdcc5fe43 req-4325e13d-307f-4519-85f0-544c0e180444 service nova] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Received event network-vif-deleted-d16b0ff3-2873-4aa9-b5da-c97c6e5d0633 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 888.140978] env[61728]: INFO nova.compute.manager [req-797cc5aa-d65e-4535-89a3-590bdcc5fe43 req-4325e13d-307f-4519-85f0-544c0e180444 service nova] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Neutron deleted interface d16b0ff3-2873-4aa9-b5da-c97c6e5d0633; detaching it from the instance and deleting it from the info cache [ 888.140978] env[61728]: DEBUG nova.network.neutron [req-797cc5aa-d65e-4535-89a3-590bdcc5fe43 req-4325e13d-307f-4519-85f0-544c0e180444 service nova] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.204790] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464404, 'name': Rename_Task, 'duration_secs': 0.171634} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.205156] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 888.205445] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e126d61-4aa2-4d1d-a5c4-939a3b6ac0e9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.211675] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 888.211675] env[61728]: value = "task-464405" [ 888.211675] env[61728]: _type = "Task" [ 888.211675] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.220215] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464405, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.359490] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "refresh_cache-5fa2e9f4-6d68-46c2-a549-95d216bab886" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.359853] env[61728]: DEBUG nova.compute.manager [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Instance network_info: |[{"id": "32361465-1ad4-4a38-9989-8110f57692f7", "address": "fa:16:3e:c3:49:72", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32361465-1a", "ovs_interfaceid": "32361465-1ad4-4a38-9989-8110f57692f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 888.360345] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:49:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2be3fdb5-359e-43bd-8c20-2ff00e81db55', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '32361465-1ad4-4a38-9989-8110f57692f7', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.368361] env[61728]: DEBUG oslo.service.loopingcall [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.368563] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 888.368872] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca52bc2c-98ff-454a-a859-769a2ae027f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.389809] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.389809] env[61728]: value = "task-464406" [ 888.389809] env[61728]: _type = "Task" [ 888.389809] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.397291] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464406, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.458710] env[61728]: DEBUG nova.compute.manager [req-afd9253a-db66-490b-b486-ed3dd85c0d11 req-d787cc8b-a792-49ba-aa4e-2855b0644b8b service nova] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Received event network-changed-32361465-1ad4-4a38-9989-8110f57692f7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 888.458920] env[61728]: DEBUG nova.compute.manager [req-afd9253a-db66-490b-b486-ed3dd85c0d11 req-d787cc8b-a792-49ba-aa4e-2855b0644b8b service nova] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Refreshing instance network info cache due to event network-changed-32361465-1ad4-4a38-9989-8110f57692f7. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 888.459162] env[61728]: DEBUG oslo_concurrency.lockutils [req-afd9253a-db66-490b-b486-ed3dd85c0d11 req-d787cc8b-a792-49ba-aa4e-2855b0644b8b service nova] Acquiring lock "refresh_cache-5fa2e9f4-6d68-46c2-a549-95d216bab886" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.459310] env[61728]: DEBUG oslo_concurrency.lockutils [req-afd9253a-db66-490b-b486-ed3dd85c0d11 req-d787cc8b-a792-49ba-aa4e-2855b0644b8b service nova] Acquired lock "refresh_cache-5fa2e9f4-6d68-46c2-a549-95d216bab886" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.459472] env[61728]: DEBUG nova.network.neutron [req-afd9253a-db66-490b-b486-ed3dd85c0d11 req-d787cc8b-a792-49ba-aa4e-2855b0644b8b service nova] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Refreshing network info cache for port 32361465-1ad4-4a38-9989-8110f57692f7 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 888.561042] env[61728]: DEBUG oslo_concurrency.lockutils [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Releasing lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.573023] env[61728]: DEBUG nova.compute.manager [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 888.601505] env[61728]: DEBUG nova.virt.hardware [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 888.601505] env[61728]: DEBUG nova.virt.hardware [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 888.601505] env[61728]: DEBUG nova.virt.hardware [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 888.601505] env[61728]: DEBUG nova.virt.hardware [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 888.601505] env[61728]: DEBUG nova.virt.hardware [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 888.601505] env[61728]: DEBUG nova.virt.hardware [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 888.601505] env[61728]: DEBUG nova.virt.hardware [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 888.601505] env[61728]: DEBUG nova.virt.hardware [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 888.601505] env[61728]: DEBUG nova.virt.hardware [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 888.601505] env[61728]: DEBUG nova.virt.hardware [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 888.601505] env[61728]: DEBUG nova.virt.hardware [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 888.602201] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dfecd55-78e8-4f06-a9ae-86b192c00c6f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.607819] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.400s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.610605] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.973s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.616017] env[61728]: INFO nova.compute.claims [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 888.623542] env[61728]: DEBUG nova.network.neutron [-] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.627325] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5854cb6e-0e67-4b47-a2f9-c79254ea427e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.645059] env[61728]: INFO nova.scheduler.client.report [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleted allocations for instance d44718bd-872b-401c-aa11-f10bea4a35d8 [ 888.649331] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b9f80797-bd9f-4b58-a1d5-9f35fa801067 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.659181] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e1cba2-aafb-440e-b513-822a273ab1ab {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.691605] env[61728]: DEBUG nova.compute.manager [req-797cc5aa-d65e-4535-89a3-590bdcc5fe43 req-4325e13d-307f-4519-85f0-544c0e180444 service nova] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Detach interface failed, port_id=d16b0ff3-2873-4aa9-b5da-c97c6e5d0633, reason: Instance 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 888.726784] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464405, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.889058] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 888.890027] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a6475c-743b-4497-bdad-2a6045d6ecd5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.901547] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464406, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.903639] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 888.903894] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd0265e8-dfbd-4286-86da-754762f94b03 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.126665] env[61728]: INFO nova.compute.manager [-] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Took 1.41 seconds to deallocate network for instance. [ 889.155454] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8475c94-fd93-4437-adc7-8a1e881780dd tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "d44718bd-872b-401c-aa11-f10bea4a35d8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.782s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.194584] env[61728]: DEBUG nova.network.neutron [req-afd9253a-db66-490b-b486-ed3dd85c0d11 req-d787cc8b-a792-49ba-aa4e-2855b0644b8b service nova] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Updated VIF entry in instance network info cache for port 32361465-1ad4-4a38-9989-8110f57692f7. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 889.194965] env[61728]: DEBUG nova.network.neutron [req-afd9253a-db66-490b-b486-ed3dd85c0d11 req-d787cc8b-a792-49ba-aa4e-2855b0644b8b service nova] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Updating instance_info_cache with network_info: [{"id": "32361465-1ad4-4a38-9989-8110f57692f7", "address": "fa:16:3e:c3:49:72", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32361465-1a", "ovs_interfaceid": "32361465-1ad4-4a38-9989-8110f57692f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.225671] env[61728]: DEBUG oslo_vmware.api [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464405, 'name': PowerOnVM_Task, 'duration_secs': 0.774032} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.225961] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 889.226304] env[61728]: INFO nova.compute.manager [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Took 8.38 seconds to spawn the instance on the hypervisor. [ 889.226516] env[61728]: DEBUG nova.compute.manager [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.227421] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8f4d46-3c6f-4968-81f7-9fc6cfaab5e3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.343831] env[61728]: DEBUG nova.network.neutron [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Successfully updated port: fbf3382b-a343-4cc4-9233-ebf4ac3139a4 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 889.403970] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464406, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.635134] env[61728]: DEBUG oslo_concurrency.lockutils [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.697824] env[61728]: DEBUG oslo_concurrency.lockutils [req-afd9253a-db66-490b-b486-ed3dd85c0d11 req-d787cc8b-a792-49ba-aa4e-2855b0644b8b service nova] Releasing lock "refresh_cache-5fa2e9f4-6d68-46c2-a549-95d216bab886" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.743060] env[61728]: INFO nova.compute.manager [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Took 45.11 seconds to build instance. [ 889.845022] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.845294] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.845408] env[61728]: DEBUG nova.network.neutron [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 889.910149] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464406, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.960330] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331ad8bd-66cc-4741-9344-e521fedc53db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.967919] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff40f2b-91d8-4801-a745-d24a977ee677 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.000383] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9989c39-fd60-46bc-8e31-5458c242b74a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.006154] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 890.006335] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 890.006455] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Deleting the datastore file [datastore1] 174607a1-9bc8-4e07-8993-7f0bb0f308e9 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.007339] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 890.007339] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 890.007339] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Deleting the datastore file [datastore1] 121bef05-ef7b-47ee-b737-15f8b9d91158 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.007339] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 890.007593] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 890.011077] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Deleting the datastore file [datastore1] 3bfefcfc-db97-4a9d-86cb-9fb1d8158863 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.011077] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-93f76337-8bc9-4679-a7d0-a406f78dc65f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.011077] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79be74d9-5b5c-4510-ba99-ae486a902e86 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.011829] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af3d2572-6096-4c3d-bc26-8ba1786bc7ed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.019193] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c5c831-bd56-49f2-bb39-425f39a226a0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.026948] env[61728]: DEBUG oslo_vmware.api [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 890.026948] env[61728]: value = "task-464408" [ 890.026948] env[61728]: _type = "Task" [ 890.026948] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.027347] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 890.027347] env[61728]: value = "task-464410" [ 890.027347] env[61728]: _type = "Task" [ 890.027347] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.027684] env[61728]: DEBUG oslo_vmware.api [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Waiting for the task: (returnval){ [ 890.027684] env[61728]: value = "task-464409" [ 890.027684] env[61728]: _type = "Task" [ 890.027684] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.044662] env[61728]: DEBUG nova.compute.provider_tree [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.053806] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464410, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.054106] env[61728]: DEBUG oslo_vmware.api [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464408, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.057687] env[61728]: DEBUG oslo_vmware.api [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464409, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.111246] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "aebc4c7c-09d6-484f-8ad6-2416b3a41828" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.111490] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "aebc4c7c-09d6-484f-8ad6-2416b3a41828" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.170637] env[61728]: DEBUG nova.compute.manager [req-62a71e98-212b-47b5-aa52-808d25f0234f req-fb52e9fe-b80c-42f2-9b9c-7dfd1de62481 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Received event network-vif-unplugged-124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 890.170941] env[61728]: DEBUG oslo_concurrency.lockutils [req-62a71e98-212b-47b5-aa52-808d25f0234f req-fb52e9fe-b80c-42f2-9b9c-7dfd1de62481 service nova] Acquiring lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.170941] env[61728]: DEBUG oslo_concurrency.lockutils [req-62a71e98-212b-47b5-aa52-808d25f0234f req-fb52e9fe-b80c-42f2-9b9c-7dfd1de62481 service nova] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.171280] env[61728]: DEBUG oslo_concurrency.lockutils [req-62a71e98-212b-47b5-aa52-808d25f0234f req-fb52e9fe-b80c-42f2-9b9c-7dfd1de62481 service nova] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.171322] env[61728]: DEBUG nova.compute.manager [req-62a71e98-212b-47b5-aa52-808d25f0234f req-fb52e9fe-b80c-42f2-9b9c-7dfd1de62481 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] No waiting events found dispatching network-vif-unplugged-124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 890.171481] env[61728]: WARNING nova.compute.manager [req-62a71e98-212b-47b5-aa52-808d25f0234f req-fb52e9fe-b80c-42f2-9b9c-7dfd1de62481 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Received unexpected event network-vif-unplugged-124d4780-15ec-4f17-a2c4-edf1517ed589 for instance with vm_state shelved and task_state shelving_offloading. [ 890.171660] env[61728]: DEBUG nova.compute.manager [req-62a71e98-212b-47b5-aa52-808d25f0234f req-fb52e9fe-b80c-42f2-9b9c-7dfd1de62481 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Received event network-changed-124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 890.171833] env[61728]: DEBUG nova.compute.manager [req-62a71e98-212b-47b5-aa52-808d25f0234f req-fb52e9fe-b80c-42f2-9b9c-7dfd1de62481 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Refreshing instance network info cache due to event network-changed-124d4780-15ec-4f17-a2c4-edf1517ed589. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 890.172039] env[61728]: DEBUG oslo_concurrency.lockutils [req-62a71e98-212b-47b5-aa52-808d25f0234f req-fb52e9fe-b80c-42f2-9b9c-7dfd1de62481 service nova] Acquiring lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.172186] env[61728]: DEBUG oslo_concurrency.lockutils [req-62a71e98-212b-47b5-aa52-808d25f0234f req-fb52e9fe-b80c-42f2-9b9c-7dfd1de62481 service nova] Acquired lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.172349] env[61728]: DEBUG nova.network.neutron [req-62a71e98-212b-47b5-aa52-808d25f0234f req-fb52e9fe-b80c-42f2-9b9c-7dfd1de62481 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Refreshing network info cache for port 124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 890.244611] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1f24af3b-2869-42c5-85fb-8d086cbf2685 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "f27f22aa-f05d-4981-a389-311731c51f93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.656s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.390178] env[61728]: DEBUG nova.network.neutron [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 890.404237] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464406, 'name': CreateVM_Task, 'duration_secs': 1.825831} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.406225] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 890.406909] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.407091] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.407408] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.407895] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f23f6b8-3ec7-4b36-9bff-30d26c00439d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.412441] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 890.412441] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f7449a-a3ef-8f7a-e5cb-6c35d45df390" [ 890.412441] env[61728]: _type = "Task" [ 890.412441] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.419782] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f7449a-a3ef-8f7a-e5cb-6c35d45df390, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.485343] env[61728]: DEBUG nova.compute.manager [req-7732b01d-5041-4ddf-82b9-1245d4126072 req-840e48c8-742b-468e-a4b6-7387bcf52edb service nova] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Received event network-vif-plugged-fbf3382b-a343-4cc4-9233-ebf4ac3139a4 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 890.485561] env[61728]: DEBUG oslo_concurrency.lockutils [req-7732b01d-5041-4ddf-82b9-1245d4126072 req-840e48c8-742b-468e-a4b6-7387bcf52edb service nova] Acquiring lock "3415687e-9b8d-42f8-9b18-93951b8ec7da-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.485774] env[61728]: DEBUG oslo_concurrency.lockutils [req-7732b01d-5041-4ddf-82b9-1245d4126072 req-840e48c8-742b-468e-a4b6-7387bcf52edb service nova] Lock "3415687e-9b8d-42f8-9b18-93951b8ec7da-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.485944] env[61728]: DEBUG oslo_concurrency.lockutils [req-7732b01d-5041-4ddf-82b9-1245d4126072 req-840e48c8-742b-468e-a4b6-7387bcf52edb service nova] Lock "3415687e-9b8d-42f8-9b18-93951b8ec7da-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.486131] env[61728]: DEBUG nova.compute.manager [req-7732b01d-5041-4ddf-82b9-1245d4126072 req-840e48c8-742b-468e-a4b6-7387bcf52edb service nova] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] No waiting events found dispatching network-vif-plugged-fbf3382b-a343-4cc4-9233-ebf4ac3139a4 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 890.486348] env[61728]: WARNING nova.compute.manager [req-7732b01d-5041-4ddf-82b9-1245d4126072 req-840e48c8-742b-468e-a4b6-7387bcf52edb service nova] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Received unexpected event network-vif-plugged-fbf3382b-a343-4cc4-9233-ebf4ac3139a4 for instance with vm_state building and task_state spawning. [ 890.486456] env[61728]: DEBUG nova.compute.manager [req-7732b01d-5041-4ddf-82b9-1245d4126072 req-840e48c8-742b-468e-a4b6-7387bcf52edb service nova] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Received event network-changed-fbf3382b-a343-4cc4-9233-ebf4ac3139a4 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 890.486611] env[61728]: DEBUG nova.compute.manager [req-7732b01d-5041-4ddf-82b9-1245d4126072 req-840e48c8-742b-468e-a4b6-7387bcf52edb service nova] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Refreshing instance network info cache due to event network-changed-fbf3382b-a343-4cc4-9233-ebf4ac3139a4. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 890.486778] env[61728]: DEBUG oslo_concurrency.lockutils [req-7732b01d-5041-4ddf-82b9-1245d4126072 req-840e48c8-742b-468e-a4b6-7387bcf52edb service nova] Acquiring lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.544435] env[61728]: DEBUG oslo_vmware.api [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464410, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1842} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.549723] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.549926] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 890.550122] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 890.552756] env[61728]: DEBUG nova.scheduler.client.report [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.555857] env[61728]: DEBUG oslo_vmware.api [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Task: {'id': task-464409, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16917} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.556096] env[61728]: DEBUG oslo_vmware.api [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464408, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170928} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.557159] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.557349] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 890.557587] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 890.557839] env[61728]: INFO nova.compute.manager [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Took 3.52 seconds to destroy the instance on the hypervisor. [ 890.558113] env[61728]: DEBUG oslo.service.loopingcall [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.558312] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.558457] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 890.558630] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 890.558915] env[61728]: INFO nova.compute.manager [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Took 3.50 seconds to destroy the instance on the hypervisor. [ 890.559104] env[61728]: DEBUG oslo.service.loopingcall [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.559489] env[61728]: DEBUG nova.compute.manager [-] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 890.559590] env[61728]: DEBUG nova.network.neutron [-] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 890.561094] env[61728]: DEBUG nova.compute.manager [-] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 890.561189] env[61728]: DEBUG nova.network.neutron [-] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 890.578211] env[61728]: INFO nova.scheduler.client.report [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Deleted allocations for instance 3bfefcfc-db97-4a9d-86cb-9fb1d8158863 [ 890.580076] env[61728]: DEBUG nova.network.neutron [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating instance_info_cache with network_info: [{"id": "fbf3382b-a343-4cc4-9233-ebf4ac3139a4", "address": "fa:16:3e:e3:49:f4", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbf3382b-a3", "ovs_interfaceid": "fbf3382b-a343-4cc4-9233-ebf4ac3139a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.747149] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.930418] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f7449a-a3ef-8f7a-e5cb-6c35d45df390, 'name': SearchDatastore_Task, 'duration_secs': 0.009435} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.930764] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.931014] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.931265] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.931420] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.931608] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.931906] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a980b433-a831-4201-96ce-d0b2d4707425 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.942012] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.942245] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 890.943017] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-655f36d7-9733-410a-b08d-da5becc038f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.950852] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 890.950852] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d82402-9dac-758c-2801-5901c35f5d97" [ 890.950852] env[61728]: _type = "Task" [ 890.950852] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.958881] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d82402-9dac-758c-2801-5901c35f5d97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.026016] env[61728]: DEBUG nova.network.neutron [req-62a71e98-212b-47b5-aa52-808d25f0234f req-fb52e9fe-b80c-42f2-9b9c-7dfd1de62481 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Updated VIF entry in instance network info cache for port 124d4780-15ec-4f17-a2c4-edf1517ed589. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 891.026450] env[61728]: DEBUG nova.network.neutron [req-62a71e98-212b-47b5-aa52-808d25f0234f req-fb52e9fe-b80c-42f2-9b9c-7dfd1de62481 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Updating instance_info_cache with network_info: [{"id": "124d4780-15ec-4f17-a2c4-edf1517ed589", "address": "fa:16:3e:6d:47:93", "network": {"id": "6a1908cc-f6aa-4dd7-970e-c8a07e1b16ee", "bridge": null, "label": "tempest-ServersNegativeTestJSON-2047209328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "01839050e56e4dbfa6e22471a3c60998", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap124d4780-15", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.060997] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.061578] env[61728]: DEBUG nova.compute.manager [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 891.064814] env[61728]: DEBUG oslo_concurrency.lockutils [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.925s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.064814] env[61728]: DEBUG nova.objects.instance [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lazy-loading 'resources' on Instance uuid b444a6cf-3c98-4308-afc1-6e760d30082d {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.084377] env[61728]: DEBUG oslo_concurrency.lockutils [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.084793] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.087265] env[61728]: DEBUG nova.compute.manager [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Instance network_info: |[{"id": "fbf3382b-a343-4cc4-9233-ebf4ac3139a4", "address": "fa:16:3e:e3:49:f4", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbf3382b-a3", "ovs_interfaceid": "fbf3382b-a343-4cc4-9233-ebf4ac3139a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 891.087265] env[61728]: DEBUG oslo_concurrency.lockutils [req-7732b01d-5041-4ddf-82b9-1245d4126072 req-840e48c8-742b-468e-a4b6-7387bcf52edb service nova] Acquired lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.087265] env[61728]: DEBUG nova.network.neutron [req-7732b01d-5041-4ddf-82b9-1245d4126072 req-840e48c8-742b-468e-a4b6-7387bcf52edb service nova] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Refreshing network info cache for port fbf3382b-a343-4cc4-9233-ebf4ac3139a4 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 891.087265] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:49:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c492f5cc-7ae0-4cab-823c-0d5dd8c60b26', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fbf3382b-a343-4cc4-9233-ebf4ac3139a4', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 891.096356] env[61728]: DEBUG oslo.service.loopingcall [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.099880] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 891.100435] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-729e1d67-dffc-4b38-a9b3-cbde6abcd4a9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.124241] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 891.124241] env[61728]: value = "task-464411" [ 891.124241] env[61728]: _type = "Task" [ 891.124241] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.133319] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464411, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.275151] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.447635] env[61728]: DEBUG nova.network.neutron [-] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.462255] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d82402-9dac-758c-2801-5901c35f5d97, 'name': SearchDatastore_Task, 'duration_secs': 0.009563} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.466806] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52e2ea90-ec12-45c7-b5cf-125b429c181e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.472026] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 891.472026] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523ae31e-13c5-e541-cc30-ff0522053560" [ 891.472026] env[61728]: _type = "Task" [ 891.472026] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.481622] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523ae31e-13c5-e541-cc30-ff0522053560, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.530244] env[61728]: DEBUG oslo_concurrency.lockutils [req-62a71e98-212b-47b5-aa52-808d25f0234f req-fb52e9fe-b80c-42f2-9b9c-7dfd1de62481 service nova] Releasing lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.573039] env[61728]: DEBUG nova.compute.utils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.577946] env[61728]: DEBUG nova.compute.manager [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 891.578256] env[61728]: DEBUG nova.network.neutron [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 891.626018] env[61728]: DEBUG nova.policy [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f5906ebe6f2a4263968b7b44bdfa026b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'df44b6731edb4629b112b5d6c5d42b1f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 891.648215] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464411, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.687362] env[61728]: DEBUG nova.network.neutron [-] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.714475] env[61728]: DEBUG nova.network.neutron [req-7732b01d-5041-4ddf-82b9-1245d4126072 req-840e48c8-742b-468e-a4b6-7387bcf52edb service nova] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updated VIF entry in instance network info cache for port fbf3382b-a343-4cc4-9233-ebf4ac3139a4. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 891.714868] env[61728]: DEBUG nova.network.neutron [req-7732b01d-5041-4ddf-82b9-1245d4126072 req-840e48c8-742b-468e-a4b6-7387bcf52edb service nova] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating instance_info_cache with network_info: [{"id": "fbf3382b-a343-4cc4-9233-ebf4ac3139a4", "address": "fa:16:3e:e3:49:f4", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbf3382b-a3", "ovs_interfaceid": "fbf3382b-a343-4cc4-9233-ebf4ac3139a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.952484] env[61728]: INFO nova.compute.manager [-] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Took 1.39 seconds to deallocate network for instance. [ 891.963032] env[61728]: DEBUG nova.network.neutron [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Successfully created port: 3a44f0c2-3108-48c4-a9b4-1c6257346641 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 891.983141] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523ae31e-13c5-e541-cc30-ff0522053560, 'name': SearchDatastore_Task, 'duration_secs': 0.010516} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.983141] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.983552] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 5fa2e9f4-6d68-46c2-a549-95d216bab886/5fa2e9f4-6d68-46c2-a549-95d216bab886.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 891.986570] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b56c203-6b23-477e-ac38-2d9cc7a5ed4d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.994639] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 891.994639] env[61728]: value = "task-464412" [ 891.994639] env[61728]: _type = "Task" [ 891.994639] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.004358] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464412, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.078971] env[61728]: DEBUG nova.compute.manager [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 892.084942] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3df6ac-8fde-4b69-9af9-78222d11a8fc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.092907] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4ddf0e-f04f-4131-be09-15a31de7362a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.127906] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab08ace-1053-4644-93af-2d0bb6e3d923 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.144757] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75239149-8498-470f-bf1e-145a9d4fd02b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.150325] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464411, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.161495] env[61728]: DEBUG nova.compute.provider_tree [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 892.189785] env[61728]: INFO nova.compute.manager [-] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Took 1.63 seconds to deallocate network for instance. [ 892.217982] env[61728]: DEBUG oslo_concurrency.lockutils [req-7732b01d-5041-4ddf-82b9-1245d4126072 req-840e48c8-742b-468e-a4b6-7387bcf52edb service nova] Releasing lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.349062] env[61728]: DEBUG nova.compute.manager [req-50aa0fda-e173-4ced-be9b-f366b376afe3 req-1137bcd5-6779-436e-9913-e984c91fd333 service nova] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Received event network-vif-deleted-58d83a36-5bb8-4f09-bd16-16d01ec692d7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 892.349431] env[61728]: DEBUG nova.compute.manager [req-50aa0fda-e173-4ced-be9b-f366b376afe3 req-1137bcd5-6779-436e-9913-e984c91fd333 service nova] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Received event network-vif-deleted-6969ab65-3bd4-4500-9908-9fe9949e1bab {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 892.460764] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.508884] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464412, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.641910] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464411, 'name': CreateVM_Task, 'duration_secs': 1.367169} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.642815] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 892.643124] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.643310] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.643705] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 892.643955] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bbd447d-c0da-49d6-ac69-04c9dec10390 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.648656] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 892.648656] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bf8854-ef1d-35c4-15da-c8eedf43c64d" [ 892.648656] env[61728]: _type = "Task" [ 892.648656] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.657256] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bf8854-ef1d-35c4-15da-c8eedf43c64d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.681813] env[61728]: ERROR nova.scheduler.client.report [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [req-781253e4-dd14-43ac-a13a-1a54896a01b7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-781253e4-dd14-43ac-a13a-1a54896a01b7"}]} [ 892.697348] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.699762] env[61728]: DEBUG nova.scheduler.client.report [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 892.715147] env[61728]: DEBUG nova.scheduler.client.report [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 892.715457] env[61728]: DEBUG nova.compute.provider_tree [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 892.729323] env[61728]: DEBUG nova.scheduler.client.report [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 892.748409] env[61728]: DEBUG nova.scheduler.client.report [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 893.007785] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464412, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.987241} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.008180] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 5fa2e9f4-6d68-46c2-a549-95d216bab886/5fa2e9f4-6d68-46c2-a549-95d216bab886.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 893.008299] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.008557] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71b41445-8ded-4abf-bd7b-3c7ea47ce58a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.017246] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 893.017246] env[61728]: value = "task-464413" [ 893.017246] env[61728]: _type = "Task" [ 893.017246] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.025898] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464413, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.066697] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.089388] env[61728]: DEBUG nova.compute.manager [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 893.113809] env[61728]: DEBUG nova.virt.hardware [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.113809] env[61728]: DEBUG nova.virt.hardware [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.114116] env[61728]: DEBUG nova.virt.hardware [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.114116] env[61728]: DEBUG nova.virt.hardware [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.114250] env[61728]: DEBUG nova.virt.hardware [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.114409] env[61728]: DEBUG nova.virt.hardware [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.114621] env[61728]: DEBUG nova.virt.hardware [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.114787] env[61728]: DEBUG nova.virt.hardware [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.114967] env[61728]: DEBUG nova.virt.hardware [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.115155] env[61728]: DEBUG nova.virt.hardware [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.115411] env[61728]: DEBUG nova.virt.hardware [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.116295] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0dbe067-50e4-4dac-be9b-7cbe3bac06be {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.120078] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d61e58-7567-43a1-ae16-85a1e8892440 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.129275] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328136e7-8830-409f-bcec-6d6f10826739 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.133927] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e8c57ca-c510-45fb-b35d-fc8d95db6d62 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.177640] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e6f8eac-ad45-4c9f-8584-79fbbacc7e9c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.188374] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bf8854-ef1d-35c4-15da-c8eedf43c64d, 'name': SearchDatastore_Task, 'duration_secs': 0.0137} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.189545] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b5738d0-4d54-4720-9cab-4c0b9a247979 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.193565] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.193821] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 893.194100] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.194260] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.194445] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 893.194694] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7a36817e-45fd-49a1-b031-d6f93d627695 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.205864] env[61728]: DEBUG nova.compute.provider_tree [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 893.210021] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 893.210021] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 893.210021] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1516cf3d-947a-468d-a591-02bb2053fb83 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.214207] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 893.214207] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5295e61c-c035-4909-6124-96553f03e25f" [ 893.214207] env[61728]: _type = "Task" [ 893.214207] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.223187] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5295e61c-c035-4909-6124-96553f03e25f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.499319] env[61728]: DEBUG nova.network.neutron [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Successfully updated port: 3a44f0c2-3108-48c4-a9b4-1c6257346641 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 893.528695] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464413, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05883} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.529076] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.529954] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328d8591-87ea-47a7-9d84-a13da9c009db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.552120] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 5fa2e9f4-6d68-46c2-a549-95d216bab886/5fa2e9f4-6d68-46c2-a549-95d216bab886.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.552620] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3b67e55-4b6f-4d8e-8ebb-39bef9ec5ec5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.572906] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 893.572906] env[61728]: value = "task-464414" [ 893.572906] env[61728]: _type = "Task" [ 893.572906] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.582560] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464414, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.724900] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5295e61c-c035-4909-6124-96553f03e25f, 'name': SearchDatastore_Task, 'duration_secs': 0.008415} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.725712] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5dded548-90c8-49b9-92f1-29d240c409c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.731032] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 893.731032] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eba803-2600-bdac-f4f6-8bc5df94ef54" [ 893.731032] env[61728]: _type = "Task" [ 893.731032] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.738547] env[61728]: DEBUG nova.scheduler.client.report [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 130 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 893.738805] env[61728]: DEBUG nova.compute.provider_tree [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 130 to 131 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 893.739092] env[61728]: DEBUG nova.compute.provider_tree [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 893.745862] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eba803-2600-bdac-f4f6-8bc5df94ef54, 'name': SearchDatastore_Task, 'duration_secs': 0.009577} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.746327] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.746578] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3415687e-9b8d-42f8-9b18-93951b8ec7da/3415687e-9b8d-42f8-9b18-93951b8ec7da.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 893.746821] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-27507a0c-713c-48d4-a62b-0a91b33873e8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.753028] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 893.753028] env[61728]: value = "task-464415" [ 893.753028] env[61728]: _type = "Task" [ 893.753028] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.761868] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464415, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.003139] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Acquiring lock "refresh_cache-fef6cb9e-d8c7-45d3-a7f6-e361ff704006" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.003139] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Acquired lock "refresh_cache-fef6cb9e-d8c7-45d3-a7f6-e361ff704006" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.003336] env[61728]: DEBUG nova.network.neutron [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 894.085991] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464414, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.248536] env[61728]: DEBUG oslo_concurrency.lockutils [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.184s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.250766] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.527s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.251048] env[61728]: DEBUG nova.objects.instance [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lazy-loading 'resources' on Instance uuid 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.263358] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464415, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450574} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.263599] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3415687e-9b8d-42f8-9b18-93951b8ec7da/3415687e-9b8d-42f8-9b18-93951b8ec7da.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 894.263720] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 894.264035] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b260987c-d432-4af1-ba14-6bbf9f48070c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.270030] env[61728]: INFO nova.scheduler.client.report [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Deleted allocations for instance b444a6cf-3c98-4308-afc1-6e760d30082d [ 894.274501] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 894.274501] env[61728]: value = "task-464416" [ 894.274501] env[61728]: _type = "Task" [ 894.274501] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.284461] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464416, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.516247] env[61728]: DEBUG nova.compute.manager [req-ce69c183-e310-42e6-9c05-417456231116 req-67a5f210-5160-4311-a066-5700d49ca83e service nova] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Received event network-vif-plugged-3a44f0c2-3108-48c4-a9b4-1c6257346641 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 894.516247] env[61728]: DEBUG oslo_concurrency.lockutils [req-ce69c183-e310-42e6-9c05-417456231116 req-67a5f210-5160-4311-a066-5700d49ca83e service nova] Acquiring lock "fef6cb9e-d8c7-45d3-a7f6-e361ff704006-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.516707] env[61728]: DEBUG oslo_concurrency.lockutils [req-ce69c183-e310-42e6-9c05-417456231116 req-67a5f210-5160-4311-a066-5700d49ca83e service nova] Lock "fef6cb9e-d8c7-45d3-a7f6-e361ff704006-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.516707] env[61728]: DEBUG oslo_concurrency.lockutils [req-ce69c183-e310-42e6-9c05-417456231116 req-67a5f210-5160-4311-a066-5700d49ca83e service nova] Lock "fef6cb9e-d8c7-45d3-a7f6-e361ff704006-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.516780] env[61728]: DEBUG nova.compute.manager [req-ce69c183-e310-42e6-9c05-417456231116 req-67a5f210-5160-4311-a066-5700d49ca83e service nova] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] No waiting events found dispatching network-vif-plugged-3a44f0c2-3108-48c4-a9b4-1c6257346641 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 894.516892] env[61728]: WARNING nova.compute.manager [req-ce69c183-e310-42e6-9c05-417456231116 req-67a5f210-5160-4311-a066-5700d49ca83e service nova] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Received unexpected event network-vif-plugged-3a44f0c2-3108-48c4-a9b4-1c6257346641 for instance with vm_state building and task_state spawning. [ 894.517066] env[61728]: DEBUG nova.compute.manager [req-ce69c183-e310-42e6-9c05-417456231116 req-67a5f210-5160-4311-a066-5700d49ca83e service nova] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Received event network-changed-3a44f0c2-3108-48c4-a9b4-1c6257346641 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 894.517228] env[61728]: DEBUG nova.compute.manager [req-ce69c183-e310-42e6-9c05-417456231116 req-67a5f210-5160-4311-a066-5700d49ca83e service nova] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Refreshing instance network info cache due to event network-changed-3a44f0c2-3108-48c4-a9b4-1c6257346641. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 894.517399] env[61728]: DEBUG oslo_concurrency.lockutils [req-ce69c183-e310-42e6-9c05-417456231116 req-67a5f210-5160-4311-a066-5700d49ca83e service nova] Acquiring lock "refresh_cache-fef6cb9e-d8c7-45d3-a7f6-e361ff704006" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.542535] env[61728]: DEBUG nova.network.neutron [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 894.584910] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464414, 'name': ReconfigVM_Task, 'duration_secs': 0.61368} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.585249] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 5fa2e9f4-6d68-46c2-a549-95d216bab886/5fa2e9f4-6d68-46c2-a549-95d216bab886.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.585892] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc9b4e16-e43f-42d5-99a1-b75de1f45fa4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.594341] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 894.594341] env[61728]: value = "task-464417" [ 894.594341] env[61728]: _type = "Task" [ 894.594341] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.600642] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464417, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.677351] env[61728]: DEBUG nova.network.neutron [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Updating instance_info_cache with network_info: [{"id": "3a44f0c2-3108-48c4-a9b4-1c6257346641", "address": "fa:16:3e:1d:80:96", "network": {"id": "06421f63-007d-432a-9f7e-1e4f2b3e97e8", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1517025311-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df44b6731edb4629b112b5d6c5d42b1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4105fae7-3de1-4a6c-a44b-39097c174699", "external-id": "nsx-vlan-transportzone-773", "segmentation_id": 773, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a44f0c2-31", "ovs_interfaceid": "3a44f0c2-3108-48c4-a9b4-1c6257346641", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.787029] env[61728]: DEBUG oslo_concurrency.lockutils [None req-037a68be-ef88-4b38-bc41-825c9422a399 tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "b444a6cf-3c98-4308-afc1-6e760d30082d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 43.789s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.791647] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464416, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061872} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.792209] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 894.792990] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14adc73-a619-479a-8f6e-99fff387d303 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.817391] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 3415687e-9b8d-42f8-9b18-93951b8ec7da/3415687e-9b8d-42f8-9b18-93951b8ec7da.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 894.820683] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d01862e-163c-46f2-8383-c8709914991f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.840822] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 894.840822] env[61728]: value = "task-464418" [ 894.840822] env[61728]: _type = "Task" [ 894.840822] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.851624] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464418, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.104417] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464417, 'name': Rename_Task, 'duration_secs': 0.388861} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.106989] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 895.107452] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e753865c-1750-4887-a97e-ff1813dae8da {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.114862] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 895.114862] env[61728]: value = "task-464419" [ 895.114862] env[61728]: _type = "Task" [ 895.114862] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.124920] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464419, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.130378] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b77924-8d0d-4e0c-ae21-942943caa5d4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.137786] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022b63dc-0f6e-40ef-99aa-c91beda9eaac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.167842] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262433c1-c47f-4f5d-ad05-a6bbfeca26e7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.175291] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0caad7-3757-4f7a-becd-e75fa418ecf5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.181188] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Releasing lock "refresh_cache-fef6cb9e-d8c7-45d3-a7f6-e361ff704006" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.181496] env[61728]: DEBUG nova.compute.manager [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Instance network_info: |[{"id": "3a44f0c2-3108-48c4-a9b4-1c6257346641", "address": "fa:16:3e:1d:80:96", "network": {"id": "06421f63-007d-432a-9f7e-1e4f2b3e97e8", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1517025311-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df44b6731edb4629b112b5d6c5d42b1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4105fae7-3de1-4a6c-a44b-39097c174699", "external-id": "nsx-vlan-transportzone-773", "segmentation_id": 773, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a44f0c2-31", "ovs_interfaceid": "3a44f0c2-3108-48c4-a9b4-1c6257346641", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 895.182087] env[61728]: DEBUG oslo_concurrency.lockutils [req-ce69c183-e310-42e6-9c05-417456231116 req-67a5f210-5160-4311-a066-5700d49ca83e service nova] Acquired lock "refresh_cache-fef6cb9e-d8c7-45d3-a7f6-e361ff704006" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.182185] env[61728]: DEBUG nova.network.neutron [req-ce69c183-e310-42e6-9c05-417456231116 req-67a5f210-5160-4311-a066-5700d49ca83e service nova] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Refreshing network info cache for port 3a44f0c2-3108-48c4-a9b4-1c6257346641 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 895.183363] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:80:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4105fae7-3de1-4a6c-a44b-39097c174699', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a44f0c2-3108-48c4-a9b4-1c6257346641', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 895.190853] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Creating folder: Project (df44b6731edb4629b112b5d6c5d42b1f). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 895.194779] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6ca9a530-9056-4b0b-aefa-2d2a96dcdde7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.205211] env[61728]: DEBUG nova.compute.provider_tree [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.214200] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Created folder: Project (df44b6731edb4629b112b5d6c5d42b1f) in parent group-v121913. [ 895.214406] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Creating folder: Instances. Parent ref: group-v122140. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 895.217225] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6fcb8300-3062-495d-afd3-2fe952f321c5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.223336] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Created folder: Instances in parent group-v122140. [ 895.223576] env[61728]: DEBUG oslo.service.loopingcall [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.223786] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 895.224336] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-79994cc3-f0ab-4531-b0f6-3999e5e2697c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.245296] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 895.245296] env[61728]: value = "task-464423" [ 895.245296] env[61728]: _type = "Task" [ 895.245296] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.253959] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464423, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.352615] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464418, 'name': ReconfigVM_Task, 'duration_secs': 0.370665} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.352939] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 3415687e-9b8d-42f8-9b18-93951b8ec7da/3415687e-9b8d-42f8-9b18-93951b8ec7da.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 895.353828] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9a237800-3f2b-4664-99bb-3656845f7dfd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.360648] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 895.360648] env[61728]: value = "task-464424" [ 895.360648] env[61728]: _type = "Task" [ 895.360648] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.370373] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464424, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.416355] env[61728]: DEBUG nova.network.neutron [req-ce69c183-e310-42e6-9c05-417456231116 req-67a5f210-5160-4311-a066-5700d49ca83e service nova] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Updated VIF entry in instance network info cache for port 3a44f0c2-3108-48c4-a9b4-1c6257346641. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 895.416729] env[61728]: DEBUG nova.network.neutron [req-ce69c183-e310-42e6-9c05-417456231116 req-67a5f210-5160-4311-a066-5700d49ca83e service nova] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Updating instance_info_cache with network_info: [{"id": "3a44f0c2-3108-48c4-a9b4-1c6257346641", "address": "fa:16:3e:1d:80:96", "network": {"id": "06421f63-007d-432a-9f7e-1e4f2b3e97e8", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1517025311-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df44b6731edb4629b112b5d6c5d42b1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4105fae7-3de1-4a6c-a44b-39097c174699", "external-id": "nsx-vlan-transportzone-773", "segmentation_id": 773, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a44f0c2-31", "ovs_interfaceid": "3a44f0c2-3108-48c4-a9b4-1c6257346641", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.624567] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464419, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.709026] env[61728]: DEBUG nova.scheduler.client.report [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.755613] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464423, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.870750] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464424, 'name': Rename_Task, 'duration_secs': 0.256916} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.871132] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 895.871439] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df8174ec-be31-4e66-94fe-763290918131 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.878732] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 895.878732] env[61728]: value = "task-464425" [ 895.878732] env[61728]: _type = "Task" [ 895.878732] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.887756] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464425, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.920081] env[61728]: DEBUG oslo_concurrency.lockutils [req-ce69c183-e310-42e6-9c05-417456231116 req-67a5f210-5160-4311-a066-5700d49ca83e service nova] Releasing lock "refresh_cache-fef6cb9e-d8c7-45d3-a7f6-e361ff704006" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.124697] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464419, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.214087] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.963s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.217071] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 39.567s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.218910] env[61728]: INFO nova.compute.claims [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 896.236048] env[61728]: INFO nova.scheduler.client.report [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Deleted allocations for instance 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c [ 896.259370] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464423, 'name': CreateVM_Task, 'duration_secs': 0.738178} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.259593] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 896.260993] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.260993] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.260993] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 896.261190] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd2fa6a9-351a-4cc3-9c2d-fd82225e13ac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.266740] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Waiting for the task: (returnval){ [ 896.266740] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521ec9fd-69d1-054c-a7b3-0abd60a2ff7c" [ 896.266740] env[61728]: _type = "Task" [ 896.266740] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.276958] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521ec9fd-69d1-054c-a7b3-0abd60a2ff7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.388929] env[61728]: DEBUG oslo_vmware.api [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464425, 'name': PowerOnVM_Task, 'duration_secs': 0.503516} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.390303] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 896.390303] env[61728]: INFO nova.compute.manager [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Took 7.82 seconds to spawn the instance on the hypervisor. [ 896.390303] env[61728]: DEBUG nova.compute.manager [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.390738] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8cfeb96-251d-4c81-a154-6f30948f8e18 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.626927] env[61728]: DEBUG oslo_vmware.api [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464419, 'name': PowerOnVM_Task, 'duration_secs': 1.311973} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.627266] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 896.627809] env[61728]: INFO nova.compute.manager [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Took 10.93 seconds to spawn the instance on the hypervisor. [ 896.627809] env[61728]: DEBUG nova.compute.manager [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.628638] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf1ce76-3ec9-4c01-85df-a9bfb73c48b7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.747258] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5dfae23-0030-44f1-aba4-34a074056348 tempest-MigrationsAdminTest-781971742 tempest-MigrationsAdminTest-781971742-project-member] Lock "9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 45.052s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.784045] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521ec9fd-69d1-054c-a7b3-0abd60a2ff7c, 'name': SearchDatastore_Task, 'duration_secs': 0.009563} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.784538] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.784917] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 896.785312] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.785592] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.785915] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.786691] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab38cf21-1ea2-437b-ac52-cfec7a4f42e2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.797016] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.797265] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 896.798178] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15223653-0514-42d8-bffb-e02c3fd018ce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.805306] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Waiting for the task: (returnval){ [ 896.805306] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ae1502-8e25-7d0a-72ae-ff0fc5f5eba3" [ 896.805306] env[61728]: _type = "Task" [ 896.805306] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.814417] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ae1502-8e25-7d0a-72ae-ff0fc5f5eba3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.907436] env[61728]: INFO nova.compute.manager [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Took 46.32 seconds to build instance. [ 897.151797] env[61728]: INFO nova.compute.manager [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Took 49.76 seconds to build instance. [ 897.319848] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ae1502-8e25-7d0a-72ae-ff0fc5f5eba3, 'name': SearchDatastore_Task, 'duration_secs': 0.011528} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.320789] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b52c9e7-a52c-4c93-940b-f8321b99df4a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.326274] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Waiting for the task: (returnval){ [ 897.326274] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523a2261-ed76-ce41-7734-20de138cca07" [ 897.326274] env[61728]: _type = "Task" [ 897.326274] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.337580] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523a2261-ed76-ce41-7734-20de138cca07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.409365] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8e2cfc65-8db7-4046-9361-ca03f5de59ca tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "3415687e-9b8d-42f8-9b18-93951b8ec7da" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.857s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.576470] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a18ba181-8dbc-48ff-bbb1-0d0916f07648 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.584956] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d9f0f3-b0f2-4d1a-a616-47b6d6dc7d1d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.618113] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e31ffd6-df4d-415c-b21e-474c28304a18 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.625332] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc98f988-1190-482a-bdae-84aa0efafd32 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.639221] env[61728]: DEBUG nova.compute.provider_tree [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.654112] env[61728]: DEBUG oslo_concurrency.lockutils [None req-63a60d0a-99e9-43ee-ad68-152642993dfa tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "5fa2e9f4-6d68-46c2-a549-95d216bab886" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.267s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.839569] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523a2261-ed76-ce41-7734-20de138cca07, 'name': SearchDatastore_Task, 'duration_secs': 0.02126} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.839569] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.839569] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] fef6cb9e-d8c7-45d3-a7f6-e361ff704006/fef6cb9e-d8c7-45d3-a7f6-e361ff704006.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 897.839828] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad2645a7-8882-42f3-a002-672ead0c5536 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.846803] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Waiting for the task: (returnval){ [ 897.846803] env[61728]: value = "task-464426" [ 897.846803] env[61728]: _type = "Task" [ 897.846803] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.855323] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464426, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.912050] env[61728]: DEBUG nova.compute.manager [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 897.997820] env[61728]: INFO nova.compute.manager [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Rescuing [ 897.998070] env[61728]: DEBUG oslo_concurrency.lockutils [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "refresh_cache-5fa2e9f4-6d68-46c2-a549-95d216bab886" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.998300] env[61728]: DEBUG oslo_concurrency.lockutils [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "refresh_cache-5fa2e9f4-6d68-46c2-a549-95d216bab886" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.998503] env[61728]: DEBUG nova.network.neutron [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 898.142880] env[61728]: DEBUG nova.scheduler.client.report [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.157502] env[61728]: DEBUG nova.compute.manager [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 898.359103] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464426, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.436570] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.449028] env[61728]: DEBUG nova.compute.manager [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Stashing vm_state: active {{(pid=61728) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 898.651764] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.651764] env[61728]: DEBUG nova.compute.manager [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 898.653628] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.589s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.653997] env[61728]: DEBUG nova.objects.instance [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Lazy-loading 'resources' on Instance uuid a9745dc8-6400-424c-bc10-1181ee128648 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.678661] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.726811] env[61728]: DEBUG nova.network.neutron [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Updating instance_info_cache with network_info: [{"id": "32361465-1ad4-4a38-9989-8110f57692f7", "address": "fa:16:3e:c3:49:72", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32361465-1a", "ovs_interfaceid": "32361465-1ad4-4a38-9989-8110f57692f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.782068] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.782405] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.782570] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.782758] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.782939] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.785281] env[61728]: INFO nova.compute.manager [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Terminating instance [ 898.787170] env[61728]: DEBUG nova.compute.manager [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 898.787375] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 898.788385] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76928382-33fe-4175-9751-15515165acb9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.796303] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 898.796537] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d9e68814-5584-4313-af87-88dfaef87312 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.802653] env[61728]: DEBUG oslo_vmware.api [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 898.802653] env[61728]: value = "task-464427" [ 898.802653] env[61728]: _type = "Task" [ 898.802653] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.810868] env[61728]: DEBUG oslo_vmware.api [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464427, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.858027] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464426, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.695418} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.858027] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] fef6cb9e-d8c7-45d3-a7f6-e361ff704006/fef6cb9e-d8c7-45d3-a7f6-e361ff704006.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 898.858027] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.858260] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7a111213-3baa-414f-bed1-173d52e28aa1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.865943] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Waiting for the task: (returnval){ [ 898.865943] env[61728]: value = "task-464428" [ 898.865943] env[61728]: _type = "Task" [ 898.865943] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.875521] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464428, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.974976] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.163870] env[61728]: DEBUG nova.compute.utils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 899.165950] env[61728]: DEBUG nova.compute.manager [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 899.165950] env[61728]: DEBUG nova.network.neutron [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 899.228652] env[61728]: DEBUG oslo_concurrency.lockutils [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "refresh_cache-5fa2e9f4-6d68-46c2-a549-95d216bab886" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.239185] env[61728]: DEBUG nova.policy [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31a2286aad9a40da9b16353541a2780f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49c64edaa37f41d38aba7bd5b1d0b47f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 899.315610] env[61728]: DEBUG oslo_vmware.api [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464427, 'name': PowerOffVM_Task, 'duration_secs': 0.251724} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.316316] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 899.316316] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 899.316316] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49a2a55b-28bd-44ef-95d6-1af61065179d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.378356] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464428, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070803} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.379052] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 899.380338] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb47700-0a8e-4253-ae41-de69eaa08dac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.406749] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] fef6cb9e-d8c7-45d3-a7f6-e361ff704006/fef6cb9e-d8c7-45d3-a7f6-e361ff704006.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.410659] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36feadbc-4149-4623-9868-9b3640224ac4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.426022] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 899.426022] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 899.426022] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Deleting the datastore file [datastore1] c84bfcd0-b145-4675-8b0a-5e8f94f65098 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 899.426789] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c661e00-bb3f-4ccd-b919-ef0469d0f028 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.433455] env[61728]: DEBUG oslo_vmware.api [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for the task: (returnval){ [ 899.433455] env[61728]: value = "task-464431" [ 899.433455] env[61728]: _type = "Task" [ 899.433455] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.434983] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Waiting for the task: (returnval){ [ 899.434983] env[61728]: value = "task-464430" [ 899.434983] env[61728]: _type = "Task" [ 899.434983] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.450772] env[61728]: DEBUG oslo_vmware.api [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464431, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.451084] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464430, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.568408] env[61728]: DEBUG nova.network.neutron [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Successfully created port: 83657468-5cac-4258-84f4-b105395d9cfa {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 899.577670] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7a320f-f22c-481a-9d56-0154094b9ffe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.589785] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c04a47b-1998-4124-b2e2-2cf1e2f9b731 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.629340] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7b6549-ed17-4173-b1d4-77fca15106cf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.637183] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729b485f-bc43-4a57-a9e6-ced93470b281 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.652139] env[61728]: DEBUG nova.compute.provider_tree [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 899.670414] env[61728]: DEBUG nova.compute.manager [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 899.775166] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 899.775166] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e5cdf3d4-7085-4613-9bfe-bc88cb606155 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.781363] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 899.781363] env[61728]: value = "task-464432" [ 899.781363] env[61728]: _type = "Task" [ 899.781363] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.789815] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464432, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.953801] env[61728]: DEBUG oslo_vmware.api [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464431, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.959302] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464430, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.172816] env[61728]: ERROR nova.scheduler.client.report [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] [req-319aa1b5-1845-42a2-ae5d-73592f4548ca] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-319aa1b5-1845-42a2-ae5d-73592f4548ca"}]} [ 900.196346] env[61728]: DEBUG nova.scheduler.client.report [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 900.212196] env[61728]: DEBUG nova.scheduler.client.report [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 900.212196] env[61728]: DEBUG nova.compute.provider_tree [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 900.225495] env[61728]: DEBUG nova.scheduler.client.report [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 900.247405] env[61728]: DEBUG nova.scheduler.client.report [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 900.293225] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464432, 'name': PowerOffVM_Task, 'duration_secs': 0.212662} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.297972] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 900.298566] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96e9365-f1b7-4664-9409-4f906257e36e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.320055] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c19e20-5717-48ac-b9aa-33904cad1de0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.355175] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 900.355175] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-355152ef-69d2-4ca1-80bc-0f8e2ccbf6ee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.364604] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 900.364604] env[61728]: value = "task-464433" [ 900.364604] env[61728]: _type = "Task" [ 900.364604] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.376537] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 900.377443] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.377948] env[61728]: DEBUG oslo_concurrency.lockutils [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.378295] env[61728]: DEBUG oslo_concurrency.lockutils [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.378604] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.381473] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4663cb21-722c-47af-a91f-c6fa2f2e23cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.398102] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.398468] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 900.399316] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f96ebf8-92f8-4398-a233-b6876254af3b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.407945] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 900.407945] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e4cb12-bd85-ae25-0533-581c64cdadc3" [ 900.407945] env[61728]: _type = "Task" [ 900.407945] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.418055] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e4cb12-bd85-ae25-0533-581c64cdadc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.445330] env[61728]: DEBUG oslo_vmware.api [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Task: {'id': task-464431, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.544641} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.450697] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 900.451080] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 900.451439] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 900.451841] env[61728]: INFO nova.compute.manager [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Took 1.66 seconds to destroy the instance on the hypervisor. [ 900.452333] env[61728]: DEBUG oslo.service.loopingcall [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.452852] env[61728]: DEBUG nova.compute.manager [-] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 900.453068] env[61728]: DEBUG nova.network.neutron [-] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 900.460704] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464430, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.638451] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4990c063-89a3-476b-90e1-c2f79306a52c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.646885] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-518ff59c-80c1-4e20-a43b-2d787021232f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.678771] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb6601e-b836-40f2-827f-3bdda829101a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.682690] env[61728]: DEBUG nova.compute.manager [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 900.690094] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef772da-201f-4d62-ac07-69887a9e8268 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.706281] env[61728]: DEBUG nova.compute.provider_tree [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 900.709554] env[61728]: DEBUG nova.virt.hardware [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 900.709794] env[61728]: DEBUG nova.virt.hardware [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 900.709992] env[61728]: DEBUG nova.virt.hardware [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 900.710273] env[61728]: DEBUG nova.virt.hardware [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 900.710442] env[61728]: DEBUG nova.virt.hardware [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 900.710599] env[61728]: DEBUG nova.virt.hardware [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 900.710827] env[61728]: DEBUG nova.virt.hardware [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 900.710999] env[61728]: DEBUG nova.virt.hardware [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 900.711192] env[61728]: DEBUG nova.virt.hardware [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 900.711364] env[61728]: DEBUG nova.virt.hardware [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 900.711545] env[61728]: DEBUG nova.virt.hardware [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 900.712368] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e7d63e-1b21-49f6-ad66-d923ecd47798 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.725129] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c88ad23a-8279-40b5-924b-49d9e0c23144 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.923889] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e4cb12-bd85-ae25-0533-581c64cdadc3, 'name': SearchDatastore_Task, 'duration_secs': 0.037678} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.924804] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1289a3ae-2e7e-4546-bfc3-65258649db68 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.931378] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 900.931378] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5296346b-e00f-4a58-694a-6292eb076ef0" [ 900.931378] env[61728]: _type = "Task" [ 900.931378] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.939702] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5296346b-e00f-4a58-694a-6292eb076ef0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.954387] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464430, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.038021] env[61728]: DEBUG nova.compute.manager [req-6f7f992d-a10b-455e-9333-61ed3b009aff req-bb066fd9-2244-4e70-9cf5-3926b1cb8cf0 service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Received event network-vif-plugged-83657468-5cac-4258-84f4-b105395d9cfa {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 901.038021] env[61728]: DEBUG oslo_concurrency.lockutils [req-6f7f992d-a10b-455e-9333-61ed3b009aff req-bb066fd9-2244-4e70-9cf5-3926b1cb8cf0 service nova] Acquiring lock "f9f07464-9f43-43fd-8895-2673861747fa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.038021] env[61728]: DEBUG oslo_concurrency.lockutils [req-6f7f992d-a10b-455e-9333-61ed3b009aff req-bb066fd9-2244-4e70-9cf5-3926b1cb8cf0 service nova] Lock "f9f07464-9f43-43fd-8895-2673861747fa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.038021] env[61728]: DEBUG oslo_concurrency.lockutils [req-6f7f992d-a10b-455e-9333-61ed3b009aff req-bb066fd9-2244-4e70-9cf5-3926b1cb8cf0 service nova] Lock "f9f07464-9f43-43fd-8895-2673861747fa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.038021] env[61728]: DEBUG nova.compute.manager [req-6f7f992d-a10b-455e-9333-61ed3b009aff req-bb066fd9-2244-4e70-9cf5-3926b1cb8cf0 service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] No waiting events found dispatching network-vif-plugged-83657468-5cac-4258-84f4-b105395d9cfa {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 901.038021] env[61728]: WARNING nova.compute.manager [req-6f7f992d-a10b-455e-9333-61ed3b009aff req-bb066fd9-2244-4e70-9cf5-3926b1cb8cf0 service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Received unexpected event network-vif-plugged-83657468-5cac-4258-84f4-b105395d9cfa for instance with vm_state building and task_state spawning. [ 901.085847] env[61728]: DEBUG nova.network.neutron [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Successfully updated port: 83657468-5cac-4258-84f4-b105395d9cfa {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 901.126734] env[61728]: DEBUG nova.compute.manager [req-37a84484-456f-4405-874d-192d2f0bba11 req-087319ea-8ba2-4947-8cb6-d786d1c72687 service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Received event network-vif-deleted-372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 901.127071] env[61728]: INFO nova.compute.manager [req-37a84484-456f-4405-874d-192d2f0bba11 req-087319ea-8ba2-4947-8cb6-d786d1c72687 service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Neutron deleted interface 372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8; detaching it from the instance and deleting it from the info cache [ 901.127191] env[61728]: DEBUG nova.network.neutron [req-37a84484-456f-4405-874d-192d2f0bba11 req-087319ea-8ba2-4947-8cb6-d786d1c72687 service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.247698] env[61728]: DEBUG nova.scheduler.client.report [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 133 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 901.247993] env[61728]: DEBUG nova.compute.provider_tree [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 133 to 134 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 901.248249] env[61728]: DEBUG nova.compute.provider_tree [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 901.442357] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5296346b-e00f-4a58-694a-6292eb076ef0, 'name': SearchDatastore_Task, 'duration_secs': 0.010909} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.442636] env[61728]: DEBUG oslo_concurrency.lockutils [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.442928] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 5fa2e9f4-6d68-46c2-a549-95d216bab886/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk. {{(pid=61728) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 901.443213] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac1e58e8-0dec-4ec2-9661-f8e46a0b5794 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.450594] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 901.450594] env[61728]: value = "task-464434" [ 901.450594] env[61728]: _type = "Task" [ 901.450594] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.453564] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464430, 'name': ReconfigVM_Task, 'duration_secs': 1.676172} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.456328] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Reconfigured VM instance instance-00000050 to attach disk [datastore1] fef6cb9e-d8c7-45d3-a7f6-e361ff704006/fef6cb9e-d8c7-45d3-a7f6-e361ff704006.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 901.456937] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5507b3bb-960a-454d-8401-5e263e62510c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.463792] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464434, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.465074] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Waiting for the task: (returnval){ [ 901.465074] env[61728]: value = "task-464435" [ 901.465074] env[61728]: _type = "Task" [ 901.465074] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.472764] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464435, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.507337] env[61728]: DEBUG nova.network.neutron [-] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.588803] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.588985] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.589174] env[61728]: DEBUG nova.network.neutron [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 901.630843] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-87fc712f-bc1b-40e5-9185-ea796a61e6c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.642621] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f06e9c52-b4bc-4712-8812-21345dd1a499 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.672271] env[61728]: DEBUG nova.compute.manager [req-37a84484-456f-4405-874d-192d2f0bba11 req-087319ea-8ba2-4947-8cb6-d786d1c72687 service nova] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Detach interface failed, port_id=372c7b2a-4d6e-42d9-b0f5-fc89cb1093e8, reason: Instance c84bfcd0-b145-4675-8b0a-5e8f94f65098 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 901.699049] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "87540771-c864-4d49-8ad3-b6b559841761" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.699338] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "87540771-c864-4d49-8ad3-b6b559841761" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.699468] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "87540771-c864-4d49-8ad3-b6b559841761-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.699661] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "87540771-c864-4d49-8ad3-b6b559841761-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.699831] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "87540771-c864-4d49-8ad3-b6b559841761-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.701966] env[61728]: INFO nova.compute.manager [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Terminating instance [ 901.703697] env[61728]: DEBUG nova.compute.manager [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 901.703909] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 901.704668] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2592b8a1-6956-4d35-83e1-1df4d8abe210 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.711909] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 901.712155] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7c5f5670-953a-4741-be97-73ef83847d4c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.717982] env[61728]: DEBUG oslo_vmware.api [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 901.717982] env[61728]: value = "task-464436" [ 901.717982] env[61728]: _type = "Task" [ 901.717982] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.726038] env[61728]: DEBUG oslo_vmware.api [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464436, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.755521] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.102s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.758929] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.462s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.759915] env[61728]: INFO nova.compute.claims [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.777902] env[61728]: INFO nova.scheduler.client.report [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Deleted allocations for instance a9745dc8-6400-424c-bc10-1181ee128648 [ 901.965420] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464434, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.974845] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464435, 'name': Rename_Task, 'duration_secs': 0.139112} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.975258] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 901.975506] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eede90cb-f6a5-4378-bab9-c382a1ca99e2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.984398] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Waiting for the task: (returnval){ [ 901.984398] env[61728]: value = "task-464437" [ 901.984398] env[61728]: _type = "Task" [ 901.984398] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.992542] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464437, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.010525] env[61728]: INFO nova.compute.manager [-] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Took 1.56 seconds to deallocate network for instance. [ 902.124331] env[61728]: DEBUG nova.network.neutron [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 902.229695] env[61728]: DEBUG oslo_vmware.api [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464436, 'name': PowerOffVM_Task, 'duration_secs': 0.215369} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.229928] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 902.230107] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 902.230376] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7b6b7362-9774-4493-8030-9c5a7eab4d45 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.268815] env[61728]: DEBUG nova.network.neutron [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance_info_cache with network_info: [{"id": "83657468-5cac-4258-84f4-b105395d9cfa", "address": "fa:16:3e:55:1a:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83657468-5c", "ovs_interfaceid": "83657468-5cac-4258-84f4-b105395d9cfa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.285463] env[61728]: DEBUG oslo_concurrency.lockutils [None req-90f127db-a1fa-4a69-a7e5-551a301f6014 tempest-ImagesNegativeTestJSON-399640235 tempest-ImagesNegativeTestJSON-399640235-project-member] Lock "a9745dc8-6400-424c-bc10-1181ee128648" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 46.327s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.294247] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 902.294486] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 902.294689] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Deleting the datastore file [datastore1] 87540771-c864-4d49-8ad3-b6b559841761 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.294970] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f0995186-cd7f-43da-b7b8-713f0ab5c86f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.303141] env[61728]: DEBUG oslo_vmware.api [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 902.303141] env[61728]: value = "task-464439" [ 902.303141] env[61728]: _type = "Task" [ 902.303141] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.311651] env[61728]: DEBUG oslo_vmware.api [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464439, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.463175] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464434, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.68274} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.463452] env[61728]: INFO nova.virt.vmwareapi.ds_util [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 5fa2e9f4-6d68-46c2-a549-95d216bab886/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk. [ 902.464284] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae2bb1d-4067-4073-90a2-fac282557393 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.489553] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 5fa2e9f4-6d68-46c2-a549-95d216bab886/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.489920] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6229b60-2b78-4f9c-9ff8-4d692f9678d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.510518] env[61728]: DEBUG oslo_vmware.api [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464437, 'name': PowerOnVM_Task, 'duration_secs': 0.461167} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.511722] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 902.511998] env[61728]: INFO nova.compute.manager [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Took 9.42 seconds to spawn the instance on the hypervisor. [ 902.512231] env[61728]: DEBUG nova.compute.manager [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 902.512555] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 902.512555] env[61728]: value = "task-464440" [ 902.512555] env[61728]: _type = "Task" [ 902.512555] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.513351] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9cfd272-4e5a-448f-81a5-d3f6e7de98bd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.519409] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.527323] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464440, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.776056] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.776544] env[61728]: DEBUG nova.compute.manager [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Instance network_info: |[{"id": "83657468-5cac-4258-84f4-b105395d9cfa", "address": "fa:16:3e:55:1a:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83657468-5c", "ovs_interfaceid": "83657468-5cac-4258-84f4-b105395d9cfa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 902.777456] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:1a:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e51ebca-e0f8-4b77-b155-4ff928eef130', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '83657468-5cac-4258-84f4-b105395d9cfa', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 902.790144] env[61728]: DEBUG oslo.service.loopingcall [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.790631] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 902.790877] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-36154b18-d63e-4922-b74a-3ec727651ebe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.822545] env[61728]: DEBUG oslo_vmware.api [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464439, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.827918] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 902.827918] env[61728]: value = "task-464441" [ 902.827918] env[61728]: _type = "Task" [ 902.827918] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.839810] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464441, 'name': CreateVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.024681] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464440, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.037993] env[61728]: INFO nova.compute.manager [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Took 50.42 seconds to build instance. [ 903.071857] env[61728]: DEBUG nova.compute.manager [req-6d0f6aad-6642-45c5-9c8f-d5aa3638e538 req-0f22ddd2-dc7c-45c1-93c9-c70f34df79dc service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Received event network-changed-83657468-5cac-4258-84f4-b105395d9cfa {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 903.072170] env[61728]: DEBUG nova.compute.manager [req-6d0f6aad-6642-45c5-9c8f-d5aa3638e538 req-0f22ddd2-dc7c-45c1-93c9-c70f34df79dc service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Refreshing instance network info cache due to event network-changed-83657468-5cac-4258-84f4-b105395d9cfa. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 903.072473] env[61728]: DEBUG oslo_concurrency.lockutils [req-6d0f6aad-6642-45c5-9c8f-d5aa3638e538 req-0f22ddd2-dc7c-45c1-93c9-c70f34df79dc service nova] Acquiring lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.072571] env[61728]: DEBUG oslo_concurrency.lockutils [req-6d0f6aad-6642-45c5-9c8f-d5aa3638e538 req-0f22ddd2-dc7c-45c1-93c9-c70f34df79dc service nova] Acquired lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.072742] env[61728]: DEBUG nova.network.neutron [req-6d0f6aad-6642-45c5-9c8f-d5aa3638e538 req-0f22ddd2-dc7c-45c1-93c9-c70f34df79dc service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Refreshing network info cache for port 83657468-5cac-4258-84f4-b105395d9cfa {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 903.139585] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76738216-58dd-4c32-9a60-c307edebcc3f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.147684] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8695b9d4-c51d-40d7-8fb4-de503c7a815f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.179747] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd928723-8f09-4476-9bdf-c1dfb8f7bc99 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.187395] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c55213-c9c5-4cbc-b956-464a1fb573fa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.201628] env[61728]: DEBUG nova.compute.provider_tree [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.322917] env[61728]: DEBUG oslo_vmware.api [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464439, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.617247} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.322917] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.322917] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 903.322917] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 903.322917] env[61728]: INFO nova.compute.manager [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Took 1.62 seconds to destroy the instance on the hypervisor. [ 903.322917] env[61728]: DEBUG oslo.service.loopingcall [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.322917] env[61728]: DEBUG nova.compute.manager [-] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 903.322917] env[61728]: DEBUG nova.network.neutron [-] [instance: 87540771-c864-4d49-8ad3-b6b559841761] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 903.338761] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464441, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.528653] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464440, 'name': ReconfigVM_Task, 'duration_secs': 0.653792} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.532118] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 5fa2e9f4-6d68-46c2-a549-95d216bab886/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 903.532118] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f91f3936-8ae6-435f-b14c-7f13579d5ac6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.559036] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d818d779-9d5a-4dd4-8233-d152345ad3a0 tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Lock "fef6cb9e-d8c7-45d3-a7f6-e361ff704006" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.973s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.563895] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69adeda9-d184-433d-b2fd-57bba0bbd3bb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.584561] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 903.584561] env[61728]: value = "task-464442" [ 903.584561] env[61728]: _type = "Task" [ 903.584561] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.594834] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464442, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.705153] env[61728]: DEBUG nova.scheduler.client.report [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.846769] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464441, 'name': CreateVM_Task, 'duration_secs': 0.757159} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.847113] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 903.848798] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.848986] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.849336] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 903.849607] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be9d6944-acb1-4d61-a2e7-dd4685448f36 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.857375] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 903.857375] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ff33d8-790a-eb6a-9e9b-2c8abe953327" [ 903.857375] env[61728]: _type = "Task" [ 903.857375] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.872173] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ff33d8-790a-eb6a-9e9b-2c8abe953327, 'name': SearchDatastore_Task, 'duration_secs': 0.009651} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.872483] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.872752] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 903.873009] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.873172] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.873356] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 903.873625] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-28662b07-f576-49a2-ba27-d5abe2b4e05c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.883090] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 903.883090] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 903.883090] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b612f68d-77c4-410b-8e6e-335209b8ad86 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.887976] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 903.887976] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aaa5d5-5d39-5d1c-1ba6-65140e74ec14" [ 903.887976] env[61728]: _type = "Task" [ 903.887976] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.896615] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aaa5d5-5d39-5d1c-1ba6-65140e74ec14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.097462] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464442, 'name': ReconfigVM_Task, 'duration_secs': 0.349086} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.097462] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 904.097462] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4a773bfc-1118-4a1e-8c14-9f552f3fd07c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.104328] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 904.104328] env[61728]: value = "task-464443" [ 904.104328] env[61728]: _type = "Task" [ 904.104328] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.113762] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464443, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.185110] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 904.185110] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 904.202772] env[61728]: DEBUG nova.network.neutron [req-6d0f6aad-6642-45c5-9c8f-d5aa3638e538 req-0f22ddd2-dc7c-45c1-93c9-c70f34df79dc service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updated VIF entry in instance network info cache for port 83657468-5cac-4258-84f4-b105395d9cfa. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 904.203171] env[61728]: DEBUG nova.network.neutron [req-6d0f6aad-6642-45c5-9c8f-d5aa3638e538 req-0f22ddd2-dc7c-45c1-93c9-c70f34df79dc service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance_info_cache with network_info: [{"id": "83657468-5cac-4258-84f4-b105395d9cfa", "address": "fa:16:3e:55:1a:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83657468-5c", "ovs_interfaceid": "83657468-5cac-4258-84f4-b105395d9cfa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.210704] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.212153] env[61728]: DEBUG nova.compute.manager [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 904.216465] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 38.262s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.216700] env[61728]: DEBUG nova.objects.instance [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'resources' on Instance uuid f28348d2-c062-497a-b374-521df51054ee {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.399152] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aaa5d5-5d39-5d1c-1ba6-65140e74ec14, 'name': SearchDatastore_Task, 'duration_secs': 0.010364} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.400122] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-251e4296-dbf2-4ccf-bfb5-c710d2657c72 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.406610] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 904.406610] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5218ecee-232c-1ab0-3307-b376e034b193" [ 904.406610] env[61728]: _type = "Task" [ 904.406610] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.414930] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5218ecee-232c-1ab0-3307-b376e034b193, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.521499] env[61728]: DEBUG nova.network.neutron [-] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.615148] env[61728]: DEBUG oslo_vmware.api [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464443, 'name': PowerOnVM_Task, 'duration_secs': 0.486459} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.615374] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 904.618269] env[61728]: DEBUG nova.compute.manager [None req-899f6c80-bf98-4a36-8b04-a35436eed960 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.619327] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccaf48fc-ec84-4ecf-91a4-6eb18b2dc2fe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.696711] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 904.696711] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Starting heal instance info cache {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 904.706419] env[61728]: DEBUG oslo_concurrency.lockutils [req-6d0f6aad-6642-45c5-9c8f-d5aa3638e538 req-0f22ddd2-dc7c-45c1-93c9-c70f34df79dc service nova] Releasing lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.723436] env[61728]: DEBUG nova.compute.utils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.728634] env[61728]: DEBUG nova.compute.manager [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 904.728634] env[61728]: DEBUG nova.network.neutron [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 904.776055] env[61728]: DEBUG nova.policy [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3baf998a00e494ba33dab6a0a2a88e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '836843da5be34c649d9a48a83e658288', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 904.920758] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5218ecee-232c-1ab0-3307-b376e034b193, 'name': SearchDatastore_Task, 'duration_secs': 0.009574} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.921118] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.921694] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] f9f07464-9f43-43fd-8895-2673861747fa/f9f07464-9f43-43fd-8895-2673861747fa.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 904.921694] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b78d795-6fde-4ae1-8410-63cfd2ae60b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.931554] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 904.931554] env[61728]: value = "task-464444" [ 904.931554] env[61728]: _type = "Task" [ 904.931554] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.945213] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464444, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.027907] env[61728]: INFO nova.compute.manager [-] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Took 1.71 seconds to deallocate network for instance. [ 905.109035] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd465c5-37aa-41b1-a2ab-1db346315a6a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.117017] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ab3b26-620d-41d0-9010-c44ae1b8824b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.164590] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fa6f111-2a6e-4532-ac37-234ffdf308d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.178096] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e6598e-45eb-4647-abaf-94864cdd967a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.194690] env[61728]: DEBUG nova.compute.provider_tree [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.229180] env[61728]: DEBUG nova.compute.manager [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 905.245408] env[61728]: DEBUG nova.network.neutron [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Successfully created port: 51b6c967-37bc-4156-a23f-91040bb1c4a2 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 905.345023] env[61728]: DEBUG nova.compute.manager [req-134acd5f-0065-4873-a341-182831a517d2 req-c0cc7389-445f-4f7c-bebc-253d2b182f4e service nova] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Received event network-vif-deleted-65322117-bd36-4d0d-b18c-dae8aac2fbba {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 905.441255] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464444, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.441539] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] f9f07464-9f43-43fd-8895-2673861747fa/f9f07464-9f43-43fd-8895-2673861747fa.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 905.441782] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 905.442061] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c322f47c-aa2c-4772-bbe0-11135f8ff1cb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.447644] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 905.447644] env[61728]: value = "task-464445" [ 905.447644] env[61728]: _type = "Task" [ 905.447644] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.457348] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464445, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.535450] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.701026] env[61728]: DEBUG nova.scheduler.client.report [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.959662] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464445, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.165695} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.960046] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 905.960865] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-366ab688-6b18-40c0-814c-6b154f2e2de9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.000342] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] f9f07464-9f43-43fd-8895-2673861747fa/f9f07464-9f43-43fd-8895-2673861747fa.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 906.001307] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db9b37f3-123c-465e-a955-a3b4a10bca2c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.036420] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 906.036420] env[61728]: value = "task-464446" [ 906.036420] env[61728]: _type = "Task" [ 906.036420] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.048523] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464446, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.205371] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.989s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.209074] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 36.528s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.210648] env[61728]: INFO nova.compute.claims [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.233908] env[61728]: INFO nova.scheduler.client.report [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Deleted allocations for instance f28348d2-c062-497a-b374-521df51054ee [ 906.241427] env[61728]: DEBUG nova.compute.manager [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 906.270474] env[61728]: DEBUG nova.virt.hardware [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.271415] env[61728]: DEBUG nova.virt.hardware [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.271737] env[61728]: DEBUG nova.virt.hardware [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.271852] env[61728]: DEBUG nova.virt.hardware [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.272040] env[61728]: DEBUG nova.virt.hardware [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.272242] env[61728]: DEBUG nova.virt.hardware [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.272490] env[61728]: DEBUG nova.virt.hardware [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.272667] env[61728]: DEBUG nova.virt.hardware [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.272849] env[61728]: DEBUG nova.virt.hardware [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.273036] env[61728]: DEBUG nova.virt.hardware [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.273226] env[61728]: DEBUG nova.virt.hardware [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.274194] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd28c14-a108-4eb7-b194-536e8259a3c5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.283667] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7dd2731-974e-4655-8b6f-23a41e078c31 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.340920] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Acquiring lock "fef6cb9e-d8c7-45d3-a7f6-e361ff704006" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.341216] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Lock "fef6cb9e-d8c7-45d3-a7f6-e361ff704006" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.341438] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Acquiring lock "fef6cb9e-d8c7-45d3-a7f6-e361ff704006-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.342026] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Lock "fef6cb9e-d8c7-45d3-a7f6-e361ff704006-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.342026] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Lock "fef6cb9e-d8c7-45d3-a7f6-e361ff704006-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.345822] env[61728]: INFO nova.compute.manager [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Terminating instance [ 906.347616] env[61728]: DEBUG nova.compute.manager [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 906.347817] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 906.348634] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7d9033-7a0b-4354-8faf-48b5d4293fe2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.357775] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 906.358094] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1ff3f4f-1599-484a-a25b-d2d7b5211dcd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.364572] env[61728]: DEBUG oslo_vmware.api [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Waiting for the task: (returnval){ [ 906.364572] env[61728]: value = "task-464447" [ 906.364572] env[61728]: _type = "Task" [ 906.364572] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.374705] env[61728]: DEBUG oslo_vmware.api [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464447, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.545837] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464446, 'name': ReconfigVM_Task, 'duration_secs': 0.279004} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.546218] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Reconfigured VM instance instance-00000051 to attach disk [datastore1] f9f07464-9f43-43fd-8895-2673861747fa/f9f07464-9f43-43fd-8895-2673861747fa.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.546866] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-317afdfc-967b-40a2-846b-7655422dc129 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.553450] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 906.553450] env[61728]: value = "task-464448" [ 906.553450] env[61728]: _type = "Task" [ 906.553450] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.565175] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464448, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.742101] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6291791a-a17d-409f-beaf-278cfafeef48 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "f28348d2-c062-497a-b374-521df51054ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 45.500s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.879847] env[61728]: DEBUG oslo_vmware.api [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464447, 'name': PowerOffVM_Task, 'duration_secs': 0.494232} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.879847] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 906.879847] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 906.879847] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d3b10aca-0c3c-4301-a0fe-c2f1fbf1e982 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.928659] env[61728]: DEBUG nova.compute.manager [req-354de1e1-f4e7-4b34-ae85-e143c014e125 req-f4ffe1ef-c76a-4d97-abd4-b11e9eae0e10 service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Received event network-vif-plugged-51b6c967-37bc-4156-a23f-91040bb1c4a2 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 906.928825] env[61728]: DEBUG oslo_concurrency.lockutils [req-354de1e1-f4e7-4b34-ae85-e143c014e125 req-f4ffe1ef-c76a-4d97-abd4-b11e9eae0e10 service nova] Acquiring lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.929158] env[61728]: DEBUG oslo_concurrency.lockutils [req-354de1e1-f4e7-4b34-ae85-e143c014e125 req-f4ffe1ef-c76a-4d97-abd4-b11e9eae0e10 service nova] Lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.929435] env[61728]: DEBUG oslo_concurrency.lockutils [req-354de1e1-f4e7-4b34-ae85-e143c014e125 req-f4ffe1ef-c76a-4d97-abd4-b11e9eae0e10 service nova] Lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.929658] env[61728]: DEBUG nova.compute.manager [req-354de1e1-f4e7-4b34-ae85-e143c014e125 req-f4ffe1ef-c76a-4d97-abd4-b11e9eae0e10 service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] No waiting events found dispatching network-vif-plugged-51b6c967-37bc-4156-a23f-91040bb1c4a2 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 906.929909] env[61728]: WARNING nova.compute.manager [req-354de1e1-f4e7-4b34-ae85-e143c014e125 req-f4ffe1ef-c76a-4d97-abd4-b11e9eae0e10 service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Received unexpected event network-vif-plugged-51b6c967-37bc-4156-a23f-91040bb1c4a2 for instance with vm_state building and task_state spawning. [ 906.945326] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 906.945766] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 906.945923] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Deleting the datastore file [datastore1] fef6cb9e-d8c7-45d3-a7f6-e361ff704006 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.946588] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d005db7-9033-4adc-a61c-d667ce0c6e3f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.954502] env[61728]: DEBUG oslo_vmware.api [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Waiting for the task: (returnval){ [ 906.954502] env[61728]: value = "task-464450" [ 906.954502] env[61728]: _type = "Task" [ 906.954502] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.964406] env[61728]: DEBUG oslo_vmware.api [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464450, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.063794] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464448, 'name': Rename_Task, 'duration_secs': 0.143539} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.064251] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 907.064549] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e03ce654-3142-4736-a664-d5e2eca15eb0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.072392] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 907.072392] env[61728]: value = "task-464451" [ 907.072392] env[61728]: _type = "Task" [ 907.072392] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.079517] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464451, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.446891] env[61728]: DEBUG nova.network.neutron [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Successfully updated port: 51b6c967-37bc-4156-a23f-91040bb1c4a2 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 907.469997] env[61728]: DEBUG oslo_vmware.api [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Task: {'id': task-464450, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138467} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.470749] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 907.470994] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 907.471207] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 907.471391] env[61728]: INFO nova.compute.manager [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Took 1.12 seconds to destroy the instance on the hypervisor. [ 907.471642] env[61728]: DEBUG oslo.service.loopingcall [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.471923] env[61728]: DEBUG nova.compute.manager [-] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 907.471981] env[61728]: DEBUG nova.network.neutron [-] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 907.561704] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df7c11e-ae61-4312-b317-98c517eb326c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.572342] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e9c94d-e06d-4219-8928-4c2f0402a62f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.585772] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464451, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.611188] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36ff61d6-724a-4ba5-92a5-dd2b2227d4c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.619098] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ec6f869-41c3-497e-bb17-9d86cbff26ea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.632608] env[61728]: DEBUG nova.compute.provider_tree [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.635524] env[61728]: DEBUG nova.compute.manager [req-fa8f85ef-cf61-4a90-a15b-3beaf986cdb8 req-78633439-65fd-4741-821d-4e9c1f73e45d service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Received event network-changed-51b6c967-37bc-4156-a23f-91040bb1c4a2 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 907.635662] env[61728]: DEBUG nova.compute.manager [req-fa8f85ef-cf61-4a90-a15b-3beaf986cdb8 req-78633439-65fd-4741-821d-4e9c1f73e45d service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Refreshing instance network info cache due to event network-changed-51b6c967-37bc-4156-a23f-91040bb1c4a2. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 907.635872] env[61728]: DEBUG oslo_concurrency.lockutils [req-fa8f85ef-cf61-4a90-a15b-3beaf986cdb8 req-78633439-65fd-4741-821d-4e9c1f73e45d service nova] Acquiring lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.636027] env[61728]: DEBUG oslo_concurrency.lockutils [req-fa8f85ef-cf61-4a90-a15b-3beaf986cdb8 req-78633439-65fd-4741-821d-4e9c1f73e45d service nova] Acquired lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.636199] env[61728]: DEBUG nova.network.neutron [req-fa8f85ef-cf61-4a90-a15b-3beaf986cdb8 req-78633439-65fd-4741-821d-4e9c1f73e45d service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Refreshing network info cache for port 51b6c967-37bc-4156-a23f-91040bb1c4a2 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 907.901949] env[61728]: INFO nova.compute.manager [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Rescuing [ 907.902689] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.902913] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.903109] env[61728]: DEBUG nova.network.neutron [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 907.949697] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.086085] env[61728]: DEBUG oslo_vmware.api [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464451, 'name': PowerOnVM_Task, 'duration_secs': 0.741599} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.086414] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 908.086590] env[61728]: INFO nova.compute.manager [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Took 7.40 seconds to spawn the instance on the hypervisor. [ 908.086774] env[61728]: DEBUG nova.compute.manager [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.087564] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0374bb98-9c26-4153-bff2-3381a31251ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.138821] env[61728]: DEBUG nova.scheduler.client.report [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 908.174983] env[61728]: DEBUG nova.network.neutron [req-fa8f85ef-cf61-4a90-a15b-3beaf986cdb8 req-78633439-65fd-4741-821d-4e9c1f73e45d service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 908.307206] env[61728]: DEBUG nova.network.neutron [req-fa8f85ef-cf61-4a90-a15b-3beaf986cdb8 req-78633439-65fd-4741-821d-4e9c1f73e45d service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.344391] env[61728]: DEBUG nova.network.neutron [-] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.589380] env[61728]: DEBUG nova.network.neutron [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Updating instance_info_cache with network_info: [{"id": "4a6313df-8741-4533-ae0e-1f469193c389", "address": "fa:16:3e:43:88:0c", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a6313df-87", "ovs_interfaceid": "4a6313df-8741-4533-ae0e-1f469193c389", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.608844] env[61728]: INFO nova.compute.manager [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Took 52.00 seconds to build instance. [ 908.648878] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.649687] env[61728]: DEBUG nova.compute.manager [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 908.654299] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.087s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.656291] env[61728]: INFO nova.compute.claims [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.809969] env[61728]: DEBUG oslo_concurrency.lockutils [req-fa8f85ef-cf61-4a90-a15b-3beaf986cdb8 req-78633439-65fd-4741-821d-4e9c1f73e45d service nova] Releasing lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.810422] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.810587] env[61728]: DEBUG nova.network.neutron [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 908.816159] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "a70d1c37-e792-4168-b7e1-9418b8cb7818" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.816387] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "a70d1c37-e792-4168-b7e1-9418b8cb7818" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.847300] env[61728]: INFO nova.compute.manager [-] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Took 1.37 seconds to deallocate network for instance. [ 909.091820] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.111340] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb060492-0dc6-4d9b-ab3b-fede79788311 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "f9f07464-9f43-43fd-8895-2673861747fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.513s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.164024] env[61728]: DEBUG nova.compute.utils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.165400] env[61728]: DEBUG nova.compute.manager [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 909.165574] env[61728]: DEBUG nova.network.neutron [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 909.223677] env[61728]: DEBUG nova.policy [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd32b5118a94042f797dea1fdf6bd05b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08367aaceba548fe93faaedf6371817d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 909.319040] env[61728]: DEBUG nova.compute.manager [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 909.349765] env[61728]: DEBUG nova.network.neutron [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 909.356575] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.573126] env[61728]: DEBUG nova.network.neutron [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Updating instance_info_cache with network_info: [{"id": "51b6c967-37bc-4156-a23f-91040bb1c4a2", "address": "fa:16:3e:7f:f1:5d", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b6c967-37", "ovs_interfaceid": "51b6c967-37bc-4156-a23f-91040bb1c4a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.615952] env[61728]: DEBUG nova.network.neutron [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Successfully created port: 54951cdd-ce11-45d8-9791-d96e06eb10f6 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 909.621714] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 909.622050] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b9f10aa-2a4a-425d-af25-b3fdca8fb3b4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.630096] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 909.630096] env[61728]: value = "task-464452" [ 909.630096] env[61728]: _type = "Task" [ 909.630096] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.638960] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464452, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.657559] env[61728]: DEBUG nova.compute.manager [req-2fb906c0-ee79-4b8c-9a0c-3cd592990a40 req-b2287a68-5a3f-449a-bf0d-c90e92c1d5a2 service nova] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Received event network-vif-deleted-3a44f0c2-3108-48c4-a9b4-1c6257346641 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 909.668983] env[61728]: DEBUG nova.compute.manager [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 909.853234] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.008622] env[61728]: DEBUG nova.compute.manager [req-bf3920ea-20c4-4bd9-a065-b85e8470fcd4 req-ed9987c4-edd3-4874-9587-fec7d2f4797f service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Received event network-changed-83657468-5cac-4258-84f4-b105395d9cfa {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 910.008853] env[61728]: DEBUG nova.compute.manager [req-bf3920ea-20c4-4bd9-a065-b85e8470fcd4 req-ed9987c4-edd3-4874-9587-fec7d2f4797f service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Refreshing instance network info cache due to event network-changed-83657468-5cac-4258-84f4-b105395d9cfa. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 910.009092] env[61728]: DEBUG oslo_concurrency.lockutils [req-bf3920ea-20c4-4bd9-a065-b85e8470fcd4 req-ed9987c4-edd3-4874-9587-fec7d2f4797f service nova] Acquiring lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.009348] env[61728]: DEBUG oslo_concurrency.lockutils [req-bf3920ea-20c4-4bd9-a065-b85e8470fcd4 req-ed9987c4-edd3-4874-9587-fec7d2f4797f service nova] Acquired lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.009514] env[61728]: DEBUG nova.network.neutron [req-bf3920ea-20c4-4bd9-a065-b85e8470fcd4 req-ed9987c4-edd3-4874-9587-fec7d2f4797f service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Refreshing network info cache for port 83657468-5cac-4258-84f4-b105395d9cfa {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 910.013487] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9406a22-e634-4d2c-9e16-a4dcb31bcb29 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.022200] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c420d63f-7d6a-4f05-bd44-973ed5c6a409 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.054316] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8178fcd7-b456-4df6-aaf1-9662678acf2b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.062212] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c88c74-9710-4460-870e-334d5359af43 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.078613] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.078971] env[61728]: DEBUG nova.compute.manager [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Instance network_info: |[{"id": "51b6c967-37bc-4156-a23f-91040bb1c4a2", "address": "fa:16:3e:7f:f1:5d", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b6c967-37", "ovs_interfaceid": "51b6c967-37bc-4156-a23f-91040bb1c4a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 910.079526] env[61728]: DEBUG nova.compute.provider_tree [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.083412] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:f1:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9732690c-bdcf-4e6f-9a32-42c196333eb8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '51b6c967-37bc-4156-a23f-91040bb1c4a2', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 910.088734] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Creating folder: Project (836843da5be34c649d9a48a83e658288). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 910.089821] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-02ad6586-48ae-4e27-8835-419237ec1ee2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.103087] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Created folder: Project (836843da5be34c649d9a48a83e658288) in parent group-v121913. [ 910.103087] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Creating folder: Instances. Parent ref: group-v122144. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 910.103087] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0f3edae5-9c65-4cfd-a8e6-0f351e9d4b31 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.111586] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Created folder: Instances in parent group-v122144. [ 910.111903] env[61728]: DEBUG oslo.service.loopingcall [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.112139] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 910.112389] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6178ae9-a2ba-4b40-9063-ec4ddd2a28db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.136275] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 910.136275] env[61728]: value = "task-464455" [ 910.136275] env[61728]: _type = "Task" [ 910.136275] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.143885] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464452, 'name': PowerOffVM_Task, 'duration_secs': 0.199747} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.144697] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 910.145941] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b08c8f-4072-4c95-85a6-6ab48739c55d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.151767] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464455, 'name': CreateVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.171325] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1594d185-57e4-4f14-abf6-3281c0d6efa5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.223220] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 910.223220] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b1f3b12-2a05-4ab2-9d3f-20a739bf485b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.228119] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 910.228119] env[61728]: value = "task-464456" [ 910.228119] env[61728]: _type = "Task" [ 910.228119] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.239254] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 910.239254] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.239519] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.239719] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.239962] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.240266] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-68f7141d-dba6-49e5-a57c-f6f7b9d11e4a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.252618] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.252731] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 910.253594] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a908d54-6ed3-42f6-ad31-319d03131251 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.263009] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 910.263009] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5252c47d-0f11-e275-fc1a-fc36ddba8960" [ 910.263009] env[61728]: _type = "Task" [ 910.263009] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.272097] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5252c47d-0f11-e275-fc1a-fc36ddba8960, 'name': SearchDatastore_Task, 'duration_secs': 0.009452} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.272868] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc53fd47-4175-4705-9ed3-266db909380b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.277988] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 910.277988] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525c40b2-0521-8808-3fd4-106b6d5f9811" [ 910.277988] env[61728]: _type = "Task" [ 910.277988] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.286259] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525c40b2-0521-8808-3fd4-106b6d5f9811, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.590610] env[61728]: DEBUG nova.scheduler.client.report [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.647410] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464455, 'name': CreateVM_Task, 'duration_secs': 0.297377} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.647581] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 910.648241] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.648416] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.648771] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.649045] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ed348f9-f5cf-4f28-b312-70dc827ebdaf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.654339] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 910.654339] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52893692-00fd-38c1-b382-b388b5a9df11" [ 910.654339] env[61728]: _type = "Task" [ 910.654339] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.662885] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52893692-00fd-38c1-b382-b388b5a9df11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.681716] env[61728]: DEBUG nova.compute.manager [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 910.703330] env[61728]: DEBUG nova.virt.hardware [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.703610] env[61728]: DEBUG nova.virt.hardware [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.703921] env[61728]: DEBUG nova.virt.hardware [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.704607] env[61728]: DEBUG nova.virt.hardware [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.704823] env[61728]: DEBUG nova.virt.hardware [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.704996] env[61728]: DEBUG nova.virt.hardware [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.705265] env[61728]: DEBUG nova.virt.hardware [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.705521] env[61728]: DEBUG nova.virt.hardware [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.705770] env[61728]: DEBUG nova.virt.hardware [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.706117] env[61728]: DEBUG nova.virt.hardware [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.706295] env[61728]: DEBUG nova.virt.hardware [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.707159] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c56b709-8f2b-459d-8d0a-78b144b635eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.715472] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c464bda-2b0b-4202-8527-28fdb4120043 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.733587] env[61728]: DEBUG nova.network.neutron [req-bf3920ea-20c4-4bd9-a065-b85e8470fcd4 req-ed9987c4-edd3-4874-9587-fec7d2f4797f service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updated VIF entry in instance network info cache for port 83657468-5cac-4258-84f4-b105395d9cfa. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 910.733962] env[61728]: DEBUG nova.network.neutron [req-bf3920ea-20c4-4bd9-a065-b85e8470fcd4 req-ed9987c4-edd3-4874-9587-fec7d2f4797f service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance_info_cache with network_info: [{"id": "83657468-5cac-4258-84f4-b105395d9cfa", "address": "fa:16:3e:55:1a:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83657468-5c", "ovs_interfaceid": "83657468-5cac-4258-84f4-b105395d9cfa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.788536] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525c40b2-0521-8808-3fd4-106b6d5f9811, 'name': SearchDatastore_Task, 'duration_secs': 0.011523} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.788824] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.789107] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] f27f22aa-f05d-4981-a389-311731c51f93/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk. {{(pid=61728) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 910.789381] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9fcb3ed3-bf11-42f9-8f0e-b3672f40916b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.797848] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 910.797848] env[61728]: value = "task-464457" [ 910.797848] env[61728]: _type = "Task" [ 910.797848] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.805933] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464457, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.099318] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.099859] env[61728]: DEBUG nova.compute.manager [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 911.102935] env[61728]: DEBUG oslo_concurrency.lockutils [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 38.235s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.103164] env[61728]: DEBUG nova.objects.instance [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lazy-loading 'resources' on Instance uuid d8657743-30f9-4f8f-aca8-d98f3c557d19 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.165353] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52893692-00fd-38c1-b382-b388b5a9df11, 'name': SearchDatastore_Task, 'duration_secs': 0.04281} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.165676] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.165924] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.166223] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.166379] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.166565] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.166852] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3ddb2d7a-547a-4433-af7a-764a9cdf5a05 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.174589] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.174778] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 911.175509] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c77574d9-d674-429b-9fa5-2b505f46f5e0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.181380] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 911.181380] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52818c39-4eb4-868e-17cc-da4fb251a67b" [ 911.181380] env[61728]: _type = "Task" [ 911.181380] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.189286] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52818c39-4eb4-868e-17cc-da4fb251a67b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.236603] env[61728]: DEBUG oslo_concurrency.lockutils [req-bf3920ea-20c4-4bd9-a065-b85e8470fcd4 req-ed9987c4-edd3-4874-9587-fec7d2f4797f service nova] Releasing lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.247421] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Didn't find any instances for network info cache update. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 911.247661] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.247856] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.248044] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.248207] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.248352] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.248499] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.248622] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61728) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 911.248859] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.252378] env[61728]: DEBUG nova.network.neutron [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Successfully updated port: 54951cdd-ce11-45d8-9791-d96e06eb10f6 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 911.308742] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464457, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.609330] env[61728]: DEBUG nova.compute.utils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 911.611078] env[61728]: DEBUG nova.compute.manager [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 911.611283] env[61728]: DEBUG nova.network.neutron [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 911.651710] env[61728]: DEBUG nova.policy [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1c8ab592fc2f4f9b9dcedb6ce343f925', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '265bb37ab35443b4a651c88fc6624c26', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 911.692141] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52818c39-4eb4-868e-17cc-da4fb251a67b, 'name': SearchDatastore_Task, 'duration_secs': 0.009593} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.695393] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b90f329f-3bc8-4fd8-97df-cd0a9f0850f8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.700505] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 911.700505] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521c47dd-f51c-6482-57af-3112de97d959" [ 911.700505] env[61728]: _type = "Task" [ 911.700505] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.708191] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521c47dd-f51c-6482-57af-3112de97d959, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.751517] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.756719] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "refresh_cache-ae25dc25-f600-471f-aaaa-f166ed691bf3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.756719] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "refresh_cache-ae25dc25-f600-471f-aaaa-f166ed691bf3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.756719] env[61728]: DEBUG nova.network.neutron [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 911.808859] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464457, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.846832} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.809549] env[61728]: INFO nova.virt.vmwareapi.ds_util [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] f27f22aa-f05d-4981-a389-311731c51f93/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk. [ 911.810642] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14a231be-3973-4e4b-906d-24cb3c4fc172 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.842019] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] f27f22aa-f05d-4981-a389-311731c51f93/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.845042] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f930c7d7-5210-466a-b6ba-09552b23f627 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.865613] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 911.865613] env[61728]: value = "task-464458" [ 911.865613] env[61728]: _type = "Task" [ 911.865613] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.877369] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464458, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.926805] env[61728]: DEBUG nova.compute.manager [req-3a035385-8ef8-4e92-b510-d8ac6bffcf4e req-ac0f9bf6-505e-4cc5-a208-41d3e20ff13a service nova] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Received event network-vif-plugged-54951cdd-ce11-45d8-9791-d96e06eb10f6 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.926929] env[61728]: DEBUG oslo_concurrency.lockutils [req-3a035385-8ef8-4e92-b510-d8ac6bffcf4e req-ac0f9bf6-505e-4cc5-a208-41d3e20ff13a service nova] Acquiring lock "ae25dc25-f600-471f-aaaa-f166ed691bf3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.927139] env[61728]: DEBUG oslo_concurrency.lockutils [req-3a035385-8ef8-4e92-b510-d8ac6bffcf4e req-ac0f9bf6-505e-4cc5-a208-41d3e20ff13a service nova] Lock "ae25dc25-f600-471f-aaaa-f166ed691bf3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.927325] env[61728]: DEBUG oslo_concurrency.lockutils [req-3a035385-8ef8-4e92-b510-d8ac6bffcf4e req-ac0f9bf6-505e-4cc5-a208-41d3e20ff13a service nova] Lock "ae25dc25-f600-471f-aaaa-f166ed691bf3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.927481] env[61728]: DEBUG nova.compute.manager [req-3a035385-8ef8-4e92-b510-d8ac6bffcf4e req-ac0f9bf6-505e-4cc5-a208-41d3e20ff13a service nova] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] No waiting events found dispatching network-vif-plugged-54951cdd-ce11-45d8-9791-d96e06eb10f6 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.927679] env[61728]: WARNING nova.compute.manager [req-3a035385-8ef8-4e92-b510-d8ac6bffcf4e req-ac0f9bf6-505e-4cc5-a208-41d3e20ff13a service nova] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Received unexpected event network-vif-plugged-54951cdd-ce11-45d8-9791-d96e06eb10f6 for instance with vm_state building and task_state spawning. [ 911.927913] env[61728]: DEBUG nova.compute.manager [req-3a035385-8ef8-4e92-b510-d8ac6bffcf4e req-ac0f9bf6-505e-4cc5-a208-41d3e20ff13a service nova] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Received event network-changed-54951cdd-ce11-45d8-9791-d96e06eb10f6 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.928262] env[61728]: DEBUG nova.compute.manager [req-3a035385-8ef8-4e92-b510-d8ac6bffcf4e req-ac0f9bf6-505e-4cc5-a208-41d3e20ff13a service nova] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Refreshing instance network info cache due to event network-changed-54951cdd-ce11-45d8-9791-d96e06eb10f6. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 911.928262] env[61728]: DEBUG oslo_concurrency.lockutils [req-3a035385-8ef8-4e92-b510-d8ac6bffcf4e req-ac0f9bf6-505e-4cc5-a208-41d3e20ff13a service nova] Acquiring lock "refresh_cache-ae25dc25-f600-471f-aaaa-f166ed691bf3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.009695] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0deea829-46f7-43e2-aa19-c6768e960a89 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.018546] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe2b74c-fab0-49c0-a564-22ed75a855b6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.050122] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e36310e-963b-4de6-8ad5-fe4bc2679c30 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.057747] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-209add87-59bf-45c2-9dc4-613cf764f30c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.071451] env[61728]: DEBUG nova.compute.provider_tree [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.115952] env[61728]: DEBUG nova.compute.manager [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 912.208815] env[61728]: DEBUG nova.network.neutron [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Successfully created port: dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 912.215476] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521c47dd-f51c-6482-57af-3112de97d959, 'name': SearchDatastore_Task, 'duration_secs': 0.015506} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.215608] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.215890] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b1fb2e3f-234a-44c7-983e-d4441b4e3cef/b1fb2e3f-234a-44c7-983e-d4441b4e3cef.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 912.216179] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-951bfac6-692b-4f98-b323-425c57d8fcd8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.224069] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 912.224069] env[61728]: value = "task-464459" [ 912.224069] env[61728]: _type = "Task" [ 912.224069] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.232143] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464459, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.285469] env[61728]: DEBUG nova.network.neutron [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 912.374588] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464458, 'name': ReconfigVM_Task, 'duration_secs': 0.502904} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.374742] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Reconfigured VM instance instance-0000004d to attach disk [datastore1] f27f22aa-f05d-4981-a389-311731c51f93/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.375592] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8034219f-4686-4449-974f-250657dc6f45 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.402150] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-548fa780-49e4-4229-a065-f15034b23b6d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.418021] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 912.418021] env[61728]: value = "task-464460" [ 912.418021] env[61728]: _type = "Task" [ 912.418021] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.426708] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464460, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.434886] env[61728]: DEBUG nova.network.neutron [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Updating instance_info_cache with network_info: [{"id": "54951cdd-ce11-45d8-9791-d96e06eb10f6", "address": "fa:16:3e:8d:70:f5", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54951cdd-ce", "ovs_interfaceid": "54951cdd-ce11-45d8-9791-d96e06eb10f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.574836] env[61728]: DEBUG nova.scheduler.client.report [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.737548] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464459, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.927768] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464460, 'name': ReconfigVM_Task, 'duration_secs': 0.147716} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.928119] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 912.928383] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb010cf1-7d47-418f-b1da-ae3b92c9e0d0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.935091] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 912.935091] env[61728]: value = "task-464461" [ 912.935091] env[61728]: _type = "Task" [ 912.935091] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.938539] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "refresh_cache-ae25dc25-f600-471f-aaaa-f166ed691bf3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.938635] env[61728]: DEBUG nova.compute.manager [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Instance network_info: |[{"id": "54951cdd-ce11-45d8-9791-d96e06eb10f6", "address": "fa:16:3e:8d:70:f5", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54951cdd-ce", "ovs_interfaceid": "54951cdd-ce11-45d8-9791-d96e06eb10f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 912.938905] env[61728]: DEBUG oslo_concurrency.lockutils [req-3a035385-8ef8-4e92-b510-d8ac6bffcf4e req-ac0f9bf6-505e-4cc5-a208-41d3e20ff13a service nova] Acquired lock "refresh_cache-ae25dc25-f600-471f-aaaa-f166ed691bf3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.939093] env[61728]: DEBUG nova.network.neutron [req-3a035385-8ef8-4e92-b510-d8ac6bffcf4e req-ac0f9bf6-505e-4cc5-a208-41d3e20ff13a service nova] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Refreshing network info cache for port 54951cdd-ce11-45d8-9791-d96e06eb10f6 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 912.940324] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8d:70:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd733acc2-07d0-479e-918c-ec8a21925389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '54951cdd-ce11-45d8-9791-d96e06eb10f6', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.947696] env[61728]: DEBUG oslo.service.loopingcall [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.951402] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 912.954961] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-26ad5474-f9ed-4f99-9a6f-01b9cb2c989e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.971008] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464461, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.977503] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.977503] env[61728]: value = "task-464462" [ 912.977503] env[61728]: _type = "Task" [ 912.977503] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.987120] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464462, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.080070] env[61728]: DEBUG oslo_concurrency.lockutils [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.977s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.084756] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.764s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.085074] env[61728]: INFO nova.compute.claims [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.109437] env[61728]: INFO nova.scheduler.client.report [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleted allocations for instance d8657743-30f9-4f8f-aca8-d98f3c557d19 [ 913.130764] env[61728]: DEBUG nova.compute.manager [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 913.163362] env[61728]: DEBUG nova.virt.hardware [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 913.163362] env[61728]: DEBUG nova.virt.hardware [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 913.163362] env[61728]: DEBUG nova.virt.hardware [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 913.163557] env[61728]: DEBUG nova.virt.hardware [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 913.163671] env[61728]: DEBUG nova.virt.hardware [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 913.163968] env[61728]: DEBUG nova.virt.hardware [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 913.164258] env[61728]: DEBUG nova.virt.hardware [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 913.164449] env[61728]: DEBUG nova.virt.hardware [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 913.164706] env[61728]: DEBUG nova.virt.hardware [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 913.164895] env[61728]: DEBUG nova.virt.hardware [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 913.166443] env[61728]: DEBUG nova.virt.hardware [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 913.166443] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-721af0b7-1803-45a5-8aee-d40128ea0842 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.179109] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f0060d9-bb7c-4486-969e-32928f93c572 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.214372] env[61728]: DEBUG nova.network.neutron [req-3a035385-8ef8-4e92-b510-d8ac6bffcf4e req-ac0f9bf6-505e-4cc5-a208-41d3e20ff13a service nova] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Updated VIF entry in instance network info cache for port 54951cdd-ce11-45d8-9791-d96e06eb10f6. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 913.214794] env[61728]: DEBUG nova.network.neutron [req-3a035385-8ef8-4e92-b510-d8ac6bffcf4e req-ac0f9bf6-505e-4cc5-a208-41d3e20ff13a service nova] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Updating instance_info_cache with network_info: [{"id": "54951cdd-ce11-45d8-9791-d96e06eb10f6", "address": "fa:16:3e:8d:70:f5", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54951cdd-ce", "ovs_interfaceid": "54951cdd-ce11-45d8-9791-d96e06eb10f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.240344] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464459, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.444818] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464461, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.486518] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464462, 'name': CreateVM_Task, 'duration_secs': 0.395863} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.486650] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 913.487407] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.487618] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.488012] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 913.488287] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d61ebccd-fbdf-4538-a4fe-fe1c3ce24ae7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.492891] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 913.492891] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5264248c-7ccd-efa6-1a2b-7794309220c6" [ 913.492891] env[61728]: _type = "Task" [ 913.492891] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.500583] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5264248c-7ccd-efa6-1a2b-7794309220c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.619309] env[61728]: DEBUG oslo_concurrency.lockutils [None req-31ca0652-0222-44b5-b5f4-acb297ec115c tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "d8657743-30f9-4f8f-aca8-d98f3c557d19" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 44.272s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.717537] env[61728]: DEBUG oslo_concurrency.lockutils [req-3a035385-8ef8-4e92-b510-d8ac6bffcf4e req-ac0f9bf6-505e-4cc5-a208-41d3e20ff13a service nova] Releasing lock "refresh_cache-ae25dc25-f600-471f-aaaa-f166ed691bf3" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.736172] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464459, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.946859] env[61728]: DEBUG oslo_vmware.api [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464461, 'name': PowerOnVM_Task, 'duration_secs': 0.664135} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.946859] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 913.949367] env[61728]: DEBUG nova.compute.manager [None req-1d850c04-a4e0-4c26-9f6f-0c862e36ab66 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 913.950252] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb901f4-8046-43f5-b577-10dade00909f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.003675] env[61728]: DEBUG nova.compute.manager [req-770f4e8e-d3a9-4abc-8513-8d71f6a31e03 req-8bc0e349-2f6d-42fe-bcc5-510bac9d1ac6 service nova] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Received event network-vif-plugged-dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 914.003969] env[61728]: DEBUG oslo_concurrency.lockutils [req-770f4e8e-d3a9-4abc-8513-8d71f6a31e03 req-8bc0e349-2f6d-42fe-bcc5-510bac9d1ac6 service nova] Acquiring lock "2400abb1-8a07-4f6b-8818-778a2105beb8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.004521] env[61728]: DEBUG oslo_concurrency.lockutils [req-770f4e8e-d3a9-4abc-8513-8d71f6a31e03 req-8bc0e349-2f6d-42fe-bcc5-510bac9d1ac6 service nova] Lock "2400abb1-8a07-4f6b-8818-778a2105beb8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.004805] env[61728]: DEBUG oslo_concurrency.lockutils [req-770f4e8e-d3a9-4abc-8513-8d71f6a31e03 req-8bc0e349-2f6d-42fe-bcc5-510bac9d1ac6 service nova] Lock "2400abb1-8a07-4f6b-8818-778a2105beb8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.005055] env[61728]: DEBUG nova.compute.manager [req-770f4e8e-d3a9-4abc-8513-8d71f6a31e03 req-8bc0e349-2f6d-42fe-bcc5-510bac9d1ac6 service nova] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] No waiting events found dispatching network-vif-plugged-dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 914.005293] env[61728]: WARNING nova.compute.manager [req-770f4e8e-d3a9-4abc-8513-8d71f6a31e03 req-8bc0e349-2f6d-42fe-bcc5-510bac9d1ac6 service nova] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Received unexpected event network-vif-plugged-dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9 for instance with vm_state building and task_state spawning. [ 914.010429] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5264248c-7ccd-efa6-1a2b-7794309220c6, 'name': SearchDatastore_Task, 'duration_secs': 0.05813} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.012040] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.012848] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.013712] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.013950] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.014571] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.017381] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ccf7c6e6-cf2a-4967-b646-901dac54e7de {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.030294] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.030294] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 914.030294] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84028b80-9ef1-42af-bc7a-e362b4262114 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.035372] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 914.035372] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528a6aa7-3428-8b57-76e4-71f21a80c75a" [ 914.035372] env[61728]: _type = "Task" [ 914.035372] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.043567] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528a6aa7-3428-8b57-76e4-71f21a80c75a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.079660] env[61728]: DEBUG nova.network.neutron [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Successfully updated port: dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 914.237621] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464459, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.393495] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe5a82b-adc6-4bcc-81de-bb00bc05c42d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.400875] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11200db6-6a71-4a20-8977-c4944d10a175 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.431783] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e3dbe6b-6e52-4055-a71d-9ff9aafea923 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.438744] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212f03c7-9594-4ac0-8b4e-bdfab972375b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.455654] env[61728]: DEBUG nova.compute.provider_tree [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.546318] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528a6aa7-3428-8b57-76e4-71f21a80c75a, 'name': SearchDatastore_Task, 'duration_secs': 0.052494} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.547188] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d5bd119-1748-4caf-88df-7065549a930e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.552269] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 914.552269] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5213b92c-5fa9-fc0e-6bc8-bfdbd7f9a208" [ 914.552269] env[61728]: _type = "Task" [ 914.552269] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.560050] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5213b92c-5fa9-fc0e-6bc8-bfdbd7f9a208, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.582806] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Acquiring lock "refresh_cache-2400abb1-8a07-4f6b-8818-778a2105beb8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.582928] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Acquired lock "refresh_cache-2400abb1-8a07-4f6b-8818-778a2105beb8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.583048] env[61728]: DEBUG nova.network.neutron [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 914.684446] env[61728]: DEBUG oslo_concurrency.lockutils [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "06109957-2c3b-404e-a43e-dd34ece39096" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.684723] env[61728]: DEBUG oslo_concurrency.lockutils [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "06109957-2c3b-404e-a43e-dd34ece39096" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.684940] env[61728]: DEBUG oslo_concurrency.lockutils [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "06109957-2c3b-404e-a43e-dd34ece39096-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.685145] env[61728]: DEBUG oslo_concurrency.lockutils [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "06109957-2c3b-404e-a43e-dd34ece39096-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.685325] env[61728]: DEBUG oslo_concurrency.lockutils [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "06109957-2c3b-404e-a43e-dd34ece39096-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.688041] env[61728]: INFO nova.compute.manager [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Terminating instance [ 914.689915] env[61728]: DEBUG nova.compute.manager [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 914.690161] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 914.690990] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb397673-121e-4984-8105-912baf2de0cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.698953] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 914.699287] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec08c5ba-474b-49c9-9cee-70b29f561930 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.705236] env[61728]: DEBUG oslo_vmware.api [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 914.705236] env[61728]: value = "task-464463" [ 914.705236] env[61728]: _type = "Task" [ 914.705236] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.713321] env[61728]: DEBUG oslo_vmware.api [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464463, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.739413] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464459, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.114124} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.739816] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b1fb2e3f-234a-44c7-983e-d4441b4e3cef/b1fb2e3f-234a-44c7-983e-d4441b4e3cef.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 914.740119] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 914.740579] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a1ff0dd9-3bd7-44a5-b336-1a69fda04d65 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.748323] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 914.748323] env[61728]: value = "task-464464" [ 914.748323] env[61728]: _type = "Task" [ 914.748323] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.757019] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464464, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.960323] env[61728]: DEBUG nova.scheduler.client.report [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 915.063407] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5213b92c-5fa9-fc0e-6bc8-bfdbd7f9a208, 'name': SearchDatastore_Task, 'duration_secs': 0.020419} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.063848] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.064019] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] ae25dc25-f600-471f-aaaa-f166ed691bf3/ae25dc25-f600-471f-aaaa-f166ed691bf3.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 915.064348] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-82c57ac7-21a6-41f0-8d78-c5c397279a06 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.072637] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 915.072637] env[61728]: value = "task-464465" [ 915.072637] env[61728]: _type = "Task" [ 915.072637] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.083134] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464465, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.130060] env[61728]: DEBUG nova.network.neutron [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 915.141880] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "0af88a53-bbe7-443e-a7dc-734e483bb730" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.141880] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "0af88a53-bbe7-443e-a7dc-734e483bb730" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.217461] env[61728]: DEBUG oslo_vmware.api [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464463, 'name': PowerOffVM_Task, 'duration_secs': 0.347985} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.218110] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 915.218316] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 915.218654] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b9aaae6b-fee8-47f1-9bd5-56bd81c68859 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.262017] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464464, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.176236} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.262017] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 915.262017] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2fb3efa-75bd-47b1-aafe-c61480146062 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.285510] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] b1fb2e3f-234a-44c7-983e-d4441b4e3cef/b1fb2e3f-234a-44c7-983e-d4441b4e3cef.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.286090] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9268d1a0-caeb-4aca-82f7-dab41d0ccdbf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.306309] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 915.306309] env[61728]: value = "task-464467" [ 915.306309] env[61728]: _type = "Task" [ 915.306309] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.313438] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464467, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.382062] env[61728]: DEBUG nova.network.neutron [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Updating instance_info_cache with network_info: [{"id": "dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9", "address": "fa:16:3e:d6:6d:30", "network": {"id": "6e0e3eec-9791-43a5-bce2-f565f42c1add", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1306405885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "265bb37ab35443b4a651c88fc6624c26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca401eaa-889a-4f9f-ac9a-56b4c41bfc06", "external-id": "nsx-vlan-transportzone-877", "segmentation_id": 877, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd1de7cf-9c", "ovs_interfaceid": "dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.466218] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.383s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.466890] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 915.470877] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.279s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.472548] env[61728]: INFO nova.compute.claims [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.484616] env[61728]: INFO nova.compute.manager [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Unrescuing [ 915.484945] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.485295] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.485461] env[61728]: DEBUG nova.network.neutron [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 915.583257] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464465, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.643390] env[61728]: DEBUG nova.compute.manager [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 915.792036] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 915.792036] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 915.792036] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleting the datastore file [datastore1] 06109957-2c3b-404e-a43e-dd34ece39096 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 915.792036] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-06686a9d-c5fd-4e66-81ef-0a48a3b00c94 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.797950] env[61728]: DEBUG oslo_vmware.api [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 915.797950] env[61728]: value = "task-464468" [ 915.797950] env[61728]: _type = "Task" [ 915.797950] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.805984] env[61728]: DEBUG oslo_vmware.api [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464468, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.816029] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464467, 'name': ReconfigVM_Task, 'duration_secs': 0.374291} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.816303] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Reconfigured VM instance instance-00000052 to attach disk [datastore1] b1fb2e3f-234a-44c7-983e-d4441b4e3cef/b1fb2e3f-234a-44c7-983e-d4441b4e3cef.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 915.816865] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4be04b29-832e-40fe-a464-d5f133914ce9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.822979] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 915.822979] env[61728]: value = "task-464469" [ 915.822979] env[61728]: _type = "Task" [ 915.822979] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.830641] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464469, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.885465] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Releasing lock "refresh_cache-2400abb1-8a07-4f6b-8818-778a2105beb8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.885827] env[61728]: DEBUG nova.compute.manager [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Instance network_info: |[{"id": "dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9", "address": "fa:16:3e:d6:6d:30", "network": {"id": "6e0e3eec-9791-43a5-bce2-f565f42c1add", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1306405885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "265bb37ab35443b4a651c88fc6624c26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca401eaa-889a-4f9f-ac9a-56b4c41bfc06", "external-id": "nsx-vlan-transportzone-877", "segmentation_id": 877, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd1de7cf-9c", "ovs_interfaceid": "dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 915.886380] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:6d:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca401eaa-889a-4f9f-ac9a-56b4c41bfc06', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.894225] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Creating folder: Project (265bb37ab35443b4a651c88fc6624c26). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 915.894620] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b5793f13-fdf9-486c-8c1a-9ef65c580da7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.904982] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Created folder: Project (265bb37ab35443b4a651c88fc6624c26) in parent group-v121913. [ 915.905273] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Creating folder: Instances. Parent ref: group-v122148. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 915.905528] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d949a093-ec32-4b7c-b175-783525b2744a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.915089] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Created folder: Instances in parent group-v122148. [ 915.915506] env[61728]: DEBUG oslo.service.loopingcall [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.915822] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 915.916155] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-168a5b0e-80ca-496d-9f0c-e4abc7c15dba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.935613] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.935613] env[61728]: value = "task-464472" [ 915.935613] env[61728]: _type = "Task" [ 915.935613] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.944513] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464472, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.977374] env[61728]: DEBUG nova.compute.utils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.978903] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 915.979082] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 916.054456] env[61728]: DEBUG nova.policy [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'df4f3ee920b04018acf4c3c70aca7e55', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a75df31158574e6182e8d7505c659f1c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 916.083117] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464465, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.722035} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.083403] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] ae25dc25-f600-471f-aaaa-f166ed691bf3/ae25dc25-f600-471f-aaaa-f166ed691bf3.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 916.083627] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 916.083890] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-294c5c41-5d94-47a3-8196-5d15d732f21b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.089988] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 916.089988] env[61728]: value = "task-464473" [ 916.089988] env[61728]: _type = "Task" [ 916.089988] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.092317] env[61728]: DEBUG nova.compute.manager [req-2e015765-f7cb-4ac1-98c1-c947b08807d6 req-5a38bfc1-5c59-4ded-a87d-5a3e1bea189d service nova] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Received event network-changed-dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 916.093825] env[61728]: DEBUG nova.compute.manager [req-2e015765-f7cb-4ac1-98c1-c947b08807d6 req-5a38bfc1-5c59-4ded-a87d-5a3e1bea189d service nova] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Refreshing instance network info cache due to event network-changed-dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 916.093825] env[61728]: DEBUG oslo_concurrency.lockutils [req-2e015765-f7cb-4ac1-98c1-c947b08807d6 req-5a38bfc1-5c59-4ded-a87d-5a3e1bea189d service nova] Acquiring lock "refresh_cache-2400abb1-8a07-4f6b-8818-778a2105beb8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.093825] env[61728]: DEBUG oslo_concurrency.lockutils [req-2e015765-f7cb-4ac1-98c1-c947b08807d6 req-5a38bfc1-5c59-4ded-a87d-5a3e1bea189d service nova] Acquired lock "refresh_cache-2400abb1-8a07-4f6b-8818-778a2105beb8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.093825] env[61728]: DEBUG nova.network.neutron [req-2e015765-f7cb-4ac1-98c1-c947b08807d6 req-5a38bfc1-5c59-4ded-a87d-5a3e1bea189d service nova] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Refreshing network info cache for port dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 916.106836] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464473, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.167925] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.232981] env[61728]: DEBUG nova.network.neutron [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Updating instance_info_cache with network_info: [{"id": "4a6313df-8741-4533-ae0e-1f469193c389", "address": "fa:16:3e:43:88:0c", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a6313df-87", "ovs_interfaceid": "4a6313df-8741-4533-ae0e-1f469193c389", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.308260] env[61728]: DEBUG oslo_vmware.api [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464468, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217444} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.308538] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 916.308721] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 916.308909] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 916.309151] env[61728]: INFO nova.compute.manager [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Took 1.62 seconds to destroy the instance on the hypervisor. [ 916.309426] env[61728]: DEBUG oslo.service.loopingcall [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.309971] env[61728]: DEBUG nova.compute.manager [-] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 916.310096] env[61728]: DEBUG nova.network.neutron [-] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 916.334118] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464469, 'name': Rename_Task, 'duration_secs': 0.158952} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.334959] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 916.334959] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee2577a1-c2bb-4171-b6d0-cbc8030ce7e4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.341903] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 916.341903] env[61728]: value = "task-464474" [ 916.341903] env[61728]: _type = "Task" [ 916.341903] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.351275] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464474, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.447911] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464472, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.485126] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 916.571030] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Successfully created port: e06b1602-d9c2-459c-bdac-2d1361b2adb9 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.600848] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464473, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.262243} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.603711] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 916.606654] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba43eadb-a3a5-4353-adc5-a9444ae6a76a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.617719] env[61728]: DEBUG nova.compute.manager [req-8515bb43-86ba-4e8c-9709-3ec3a0f192cc req-a5af0fe2-a9d0-40c7-b669-c2ab3a58a605 service nova] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Received event network-vif-deleted-4758105c-971a-4c28-a179-7d9e20f05173 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 916.617719] env[61728]: INFO nova.compute.manager [req-8515bb43-86ba-4e8c-9709-3ec3a0f192cc req-a5af0fe2-a9d0-40c7-b669-c2ab3a58a605 service nova] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Neutron deleted interface 4758105c-971a-4c28-a179-7d9e20f05173; detaching it from the instance and deleting it from the info cache [ 916.618686] env[61728]: DEBUG nova.network.neutron [req-8515bb43-86ba-4e8c-9709-3ec3a0f192cc req-a5af0fe2-a9d0-40c7-b669-c2ab3a58a605 service nova] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.647063] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] ae25dc25-f600-471f-aaaa-f166ed691bf3/ae25dc25-f600-471f-aaaa-f166ed691bf3.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.653523] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5aa87493-23a8-4270-8407-37afdb6841ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.677316] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 916.677316] env[61728]: value = "task-464475" [ 916.677316] env[61728]: _type = "Task" [ 916.677316] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.687905] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464475, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.739499] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.740195] env[61728]: DEBUG nova.objects.instance [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lazy-loading 'flavor' on Instance uuid f27f22aa-f05d-4981-a389-311731c51f93 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.855499] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464474, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.865842] env[61728]: DEBUG nova.network.neutron [req-2e015765-f7cb-4ac1-98c1-c947b08807d6 req-5a38bfc1-5c59-4ded-a87d-5a3e1bea189d service nova] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Updated VIF entry in instance network info cache for port dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 916.866154] env[61728]: DEBUG nova.network.neutron [req-2e015765-f7cb-4ac1-98c1-c947b08807d6 req-5a38bfc1-5c59-4ded-a87d-5a3e1bea189d service nova] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Updating instance_info_cache with network_info: [{"id": "dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9", "address": "fa:16:3e:d6:6d:30", "network": {"id": "6e0e3eec-9791-43a5-bce2-f565f42c1add", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1306405885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "265bb37ab35443b4a651c88fc6624c26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca401eaa-889a-4f9f-ac9a-56b4c41bfc06", "external-id": "nsx-vlan-transportzone-877", "segmentation_id": 877, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd1de7cf-9c", "ovs_interfaceid": "dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.948275] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464472, 'name': CreateVM_Task, 'duration_secs': 0.762061} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.948832] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 916.949909] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.950275] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.951242] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 916.955618] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be5740f6-09af-4a13-a39f-a7b02f17a9e5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.961775] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Waiting for the task: (returnval){ [ 916.961775] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a33f9b-8991-f847-df24-e862c750afa0" [ 916.961775] env[61728]: _type = "Task" [ 916.961775] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.973051] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a33f9b-8991-f847-df24-e862c750afa0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.974778] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e7300b4-033d-495b-9ae7-1e59d20b604d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.981419] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ac0283d-7aa0-4888-96da-898945004ada {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.020019] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c83d8e7-3bd0-40ce-9c02-972c6dd5d812 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.025903] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37763b10-934c-44f9-9d39-f4368eb9dd4f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.039746] env[61728]: DEBUG nova.compute.provider_tree [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.071639] env[61728]: DEBUG nova.network.neutron [-] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.124090] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-98c73fa7-a452-44c3-aca5-0cced892d65a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.133493] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f2d6e0-833d-433b-be65-2e1c84ea9a75 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.163843] env[61728]: DEBUG nova.compute.manager [req-8515bb43-86ba-4e8c-9709-3ec3a0f192cc req-a5af0fe2-a9d0-40c7-b669-c2ab3a58a605 service nova] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Detach interface failed, port_id=4758105c-971a-4c28-a179-7d9e20f05173, reason: Instance 06109957-2c3b-404e-a43e-dd34ece39096 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 917.185915] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464475, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.249789] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b68b2731-78dd-410d-9371-bbbaaf72ceb2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.271593] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 917.271928] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd09681d-d085-4754-9644-e1474f1c9fb8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.281077] env[61728]: DEBUG oslo_vmware.api [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 917.281077] env[61728]: value = "task-464476" [ 917.281077] env[61728]: _type = "Task" [ 917.281077] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.289625] env[61728]: DEBUG oslo_vmware.api [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464476, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.351768] env[61728]: DEBUG oslo_vmware.api [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464474, 'name': PowerOnVM_Task, 'duration_secs': 0.619845} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.352180] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 917.352291] env[61728]: INFO nova.compute.manager [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Took 11.11 seconds to spawn the instance on the hypervisor. [ 917.352461] env[61728]: DEBUG nova.compute.manager [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.353233] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771f3af9-20fd-404e-bc93-af4f506e428e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.371649] env[61728]: DEBUG oslo_concurrency.lockutils [req-2e015765-f7cb-4ac1-98c1-c947b08807d6 req-5a38bfc1-5c59-4ded-a87d-5a3e1bea189d service nova] Releasing lock "refresh_cache-2400abb1-8a07-4f6b-8818-778a2105beb8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.472216] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a33f9b-8991-f847-df24-e862c750afa0, 'name': SearchDatastore_Task, 'duration_secs': 0.010111} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.472555] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.472834] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 917.473103] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.473264] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.473452] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.473724] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51400212-959c-414a-88f0-ed6bcd5bccd3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.481809] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.482017] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 917.482741] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b83cf4fa-f103-4d42-b94e-ae77af1af55f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.487844] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Waiting for the task: (returnval){ [ 917.487844] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ab59b2-989a-9b40-c74e-4d4e76fc59c2" [ 917.487844] env[61728]: _type = "Task" [ 917.487844] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.495055] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ab59b2-989a-9b40-c74e-4d4e76fc59c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.519493] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 917.546030] env[61728]: DEBUG nova.scheduler.client.report [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.549711] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.549852] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.550748] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.550866] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.551079] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.552248] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.552248] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.552248] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.552248] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.552248] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.552248] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.554246] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e1cd21-7d7f-4d36-8878-e43489180fc1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.561887] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52982415-c1de-41d9-811e-73772c0c910a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.575418] env[61728]: INFO nova.compute.manager [-] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Took 1.27 seconds to deallocate network for instance. [ 917.688050] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464475, 'name': ReconfigVM_Task, 'duration_secs': 0.517838} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.688362] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Reconfigured VM instance instance-00000053 to attach disk [datastore1] ae25dc25-f600-471f-aaaa-f166ed691bf3/ae25dc25-f600-471f-aaaa-f166ed691bf3.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.689016] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-16b08183-5501-4b7f-b1f0-eed450e1ab04 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.695215] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 917.695215] env[61728]: value = "task-464477" [ 917.695215] env[61728]: _type = "Task" [ 917.695215] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.702848] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464477, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.792202] env[61728]: DEBUG oslo_vmware.api [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464476, 'name': PowerOffVM_Task, 'duration_secs': 0.509269} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.792485] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 917.797782] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 917.798065] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25ecf3c7-63ee-4a73-8165-e505cd91cd35 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.816458] env[61728]: DEBUG oslo_vmware.api [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 917.816458] env[61728]: value = "task-464478" [ 917.816458] env[61728]: _type = "Task" [ 917.816458] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.824232] env[61728]: DEBUG oslo_vmware.api [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464478, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.870485] env[61728]: INFO nova.compute.manager [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Took 56.61 seconds to build instance. [ 917.999417] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ab59b2-989a-9b40-c74e-4d4e76fc59c2, 'name': SearchDatastore_Task, 'duration_secs': 0.008431} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.000291] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a9ae8e1-e7f7-470f-b0e4-bc6cf9433471 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.005663] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Waiting for the task: (returnval){ [ 918.005663] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a0810f-6794-3ed4-46b0-dc76c4360352" [ 918.005663] env[61728]: _type = "Task" [ 918.005663] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.016317] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a0810f-6794-3ed4-46b0-dc76c4360352, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.057734] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.059040] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 918.062698] env[61728]: DEBUG oslo_concurrency.lockutils [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.427s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.062698] env[61728]: DEBUG nova.objects.instance [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lazy-loading 'resources' on Instance uuid 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.081408] env[61728]: DEBUG oslo_concurrency.lockutils [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.205244] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464477, 'name': Rename_Task, 'duration_secs': 0.207554} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.205810] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 918.206114] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-41811fad-4f89-4a73-8bc3-20f3d1d74d71 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.212661] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 918.212661] env[61728]: value = "task-464479" [ 918.212661] env[61728]: _type = "Task" [ 918.212661] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.220300] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464479, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.269869] env[61728]: DEBUG nova.compute.manager [req-22e9bd5d-7dfa-4811-8c37-a664cf66d8e8 req-88e314db-c839-4346-9f44-dd725e430c0b service nova] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Received event network-vif-plugged-e06b1602-d9c2-459c-bdac-2d1361b2adb9 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 918.270122] env[61728]: DEBUG oslo_concurrency.lockutils [req-22e9bd5d-7dfa-4811-8c37-a664cf66d8e8 req-88e314db-c839-4346-9f44-dd725e430c0b service nova] Acquiring lock "0e77ef23-3d1a-4f00-83f9-5b617334f980-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.270342] env[61728]: DEBUG oslo_concurrency.lockutils [req-22e9bd5d-7dfa-4811-8c37-a664cf66d8e8 req-88e314db-c839-4346-9f44-dd725e430c0b service nova] Lock "0e77ef23-3d1a-4f00-83f9-5b617334f980-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.270566] env[61728]: DEBUG oslo_concurrency.lockutils [req-22e9bd5d-7dfa-4811-8c37-a664cf66d8e8 req-88e314db-c839-4346-9f44-dd725e430c0b service nova] Lock "0e77ef23-3d1a-4f00-83f9-5b617334f980-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.270693] env[61728]: DEBUG nova.compute.manager [req-22e9bd5d-7dfa-4811-8c37-a664cf66d8e8 req-88e314db-c839-4346-9f44-dd725e430c0b service nova] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] No waiting events found dispatching network-vif-plugged-e06b1602-d9c2-459c-bdac-2d1361b2adb9 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.270869] env[61728]: WARNING nova.compute.manager [req-22e9bd5d-7dfa-4811-8c37-a664cf66d8e8 req-88e314db-c839-4346-9f44-dd725e430c0b service nova] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Received unexpected event network-vif-plugged-e06b1602-d9c2-459c-bdac-2d1361b2adb9 for instance with vm_state building and task_state spawning. [ 918.326531] env[61728]: DEBUG oslo_vmware.api [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464478, 'name': ReconfigVM_Task, 'duration_secs': 0.304728} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.326854] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 918.327070] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 918.327337] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d49e67d2-5067-4e40-b513-6fbe671e8887 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.334406] env[61728]: DEBUG oslo_vmware.api [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 918.334406] env[61728]: value = "task-464480" [ 918.334406] env[61728]: _type = "Task" [ 918.334406] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.342232] env[61728]: DEBUG oslo_vmware.api [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464480, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.372771] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02b802ca-8ff0-4994-a48a-5ad649e5ca96 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.135s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.395096] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Successfully updated port: e06b1602-d9c2-459c-bdac-2d1361b2adb9 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.515878] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a0810f-6794-3ed4-46b0-dc76c4360352, 'name': SearchDatastore_Task, 'duration_secs': 0.008721} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.516368] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.516473] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2400abb1-8a07-4f6b-8818-778a2105beb8/2400abb1-8a07-4f6b-8818-778a2105beb8.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 918.516729] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-165a5000-6f31-4b64-85bc-c13f6e24dfa5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.524019] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Waiting for the task: (returnval){ [ 918.524019] env[61728]: value = "task-464481" [ 918.524019] env[61728]: _type = "Task" [ 918.524019] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.533746] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464481, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.564936] env[61728]: DEBUG nova.compute.utils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.569968] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 918.569968] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 918.658067] env[61728]: DEBUG nova.policy [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'df4f3ee920b04018acf4c3c70aca7e55', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a75df31158574e6182e8d7505c659f1c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 918.729754] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464479, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.849165] env[61728]: DEBUG oslo_vmware.api [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464480, 'name': PowerOnVM_Task, 'duration_secs': 0.432058} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.852325] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 918.852664] env[61728]: DEBUG nova.compute.manager [None req-eb63044b-3737-4131-828c-318de669c278 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.853862] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac430b0-cb81-4413-9d8f-c64b835f9906 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.900161] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "refresh_cache-0e77ef23-3d1a-4f00-83f9-5b617334f980" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.900161] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquired lock "refresh_cache-0e77ef23-3d1a-4f00-83f9-5b617334f980" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.900161] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 918.991450] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6fa408-5609-4afa-a20c-c666c4a36e4c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.001281] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb59cf75-19c5-40a0-b46f-9629a4207c1f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.038286] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a77dae-c824-4b12-aa35-000f0c1c6aa2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.050964] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464481, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.055895] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900ec93b-a0e4-4961-b3e7-4929da16058b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.072607] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 919.076114] env[61728]: DEBUG nova.compute.provider_tree [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.224232] env[61728]: DEBUG oslo_vmware.api [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464479, 'name': PowerOnVM_Task, 'duration_secs': 0.550703} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.224596] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 919.224848] env[61728]: INFO nova.compute.manager [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Took 8.54 seconds to spawn the instance on the hypervisor. [ 919.225110] env[61728]: DEBUG nova.compute.manager [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.225918] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a9510fe-fd6a-45f2-ae24-0819f60e8ee6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.435456] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 919.545722] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464481, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.551671} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.546017] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2400abb1-8a07-4f6b-8818-778a2105beb8/2400abb1-8a07-4f6b-8818-778a2105beb8.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 919.546247] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 919.546506] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5498e486-8b62-4284-934a-88581d0f7adf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.557913] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Waiting for the task: (returnval){ [ 919.557913] env[61728]: value = "task-464482" [ 919.557913] env[61728]: _type = "Task" [ 919.557913] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.565748] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464482, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.582254] env[61728]: DEBUG nova.scheduler.client.report [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.600921] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Updating instance_info_cache with network_info: [{"id": "e06b1602-d9c2-459c-bdac-2d1361b2adb9", "address": "fa:16:3e:60:91:58", "network": {"id": "02d8bb39-8b1a-4b90-8d0c-2873b2283fb5", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1139357359-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a75df31158574e6182e8d7505c659f1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16e15a36-a55b-4c27-b864-f284339009d0", "external-id": "nsx-vlan-transportzone-616", "segmentation_id": 616, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape06b1602-d9", "ovs_interfaceid": "e06b1602-d9c2-459c-bdac-2d1361b2adb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.704917] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Successfully created port: bcd321c5-6557-4c39-a5e4-025a4b20d7db {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 919.749139] env[61728]: INFO nova.compute.manager [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Took 50.09 seconds to build instance. [ 920.065881] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464482, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070065} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.066104] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 920.066861] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c5e4301-22f1-46a0-9436-a875b4d53112 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.088724] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 2400abb1-8a07-4f6b-8818-778a2105beb8/2400abb1-8a07-4f6b-8818-778a2105beb8.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.090012] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 920.092448] env[61728]: DEBUG oslo_concurrency.lockutils [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.030s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.094200] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd335ddc-fa96-4e25-ba2f-a29a828f1dac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.117343] env[61728]: DEBUG oslo_concurrency.lockutils [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.033s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.119408] env[61728]: DEBUG nova.objects.instance [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lazy-loading 'resources' on Instance uuid 3bfefcfc-db97-4a9d-86cb-9fb1d8158863 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.119408] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Releasing lock "refresh_cache-0e77ef23-3d1a-4f00-83f9-5b617334f980" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.119408] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Instance network_info: |[{"id": "e06b1602-d9c2-459c-bdac-2d1361b2adb9", "address": "fa:16:3e:60:91:58", "network": {"id": "02d8bb39-8b1a-4b90-8d0c-2873b2283fb5", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1139357359-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a75df31158574e6182e8d7505c659f1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16e15a36-a55b-4c27-b864-f284339009d0", "external-id": "nsx-vlan-transportzone-616", "segmentation_id": 616, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape06b1602-d9", "ovs_interfaceid": "e06b1602-d9c2-459c-bdac-2d1361b2adb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 920.120119] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:91:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '16e15a36-a55b-4c27-b864-f284339009d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e06b1602-d9c2-459c-bdac-2d1361b2adb9', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 920.129044] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Creating folder: Project (a75df31158574e6182e8d7505c659f1c). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 920.129901] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-20f88e93-c749-475c-acbf-c049c8d46939 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.133970] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Waiting for the task: (returnval){ [ 920.133970] env[61728]: value = "task-464483" [ 920.133970] env[61728]: _type = "Task" [ 920.133970] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.141657] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Created folder: Project (a75df31158574e6182e8d7505c659f1c) in parent group-v121913. [ 920.141657] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Creating folder: Instances. Parent ref: group-v122151. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 920.143492] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-921c64bc-22e4-4b69-8926-bdc7ce4a083e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.144871] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464483, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.154390] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 920.154861] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 920.154861] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 920.155155] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 920.155290] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 920.155618] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 920.159087] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 920.159087] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 920.159087] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 920.159087] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 920.159087] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.159087] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4b1bb2-3db5-436d-abd6-32eb5b480c70 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.161695] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Created folder: Instances in parent group-v122151. [ 920.161964] env[61728]: DEBUG oslo.service.loopingcall [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.163066] env[61728]: INFO nova.scheduler.client.report [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Deleted allocations for instance 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae [ 920.166344] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 920.167753] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6879b7e0-2552-4fbd-b6c9-9591997f6f29 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.189074] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a12cb2a-300e-4181-acbb-50428c5aa04d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.194390] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 920.194390] env[61728]: value = "task-464486" [ 920.194390] env[61728]: _type = "Task" [ 920.194390] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.223296] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464486, 'name': CreateVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.254663] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e695d916-18de-4344-81eb-334dee5b0bc8 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "ae25dc25-f600-471f-aaaa-f166ed691bf3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.406s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.406911] env[61728]: DEBUG nova.compute.manager [req-0a2746f3-0be6-4cf5-9177-3eb8f68ebc9d req-f5e71ef9-e4b8-43af-bfaf-d8194ef85eee service nova] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Received event network-changed-e06b1602-d9c2-459c-bdac-2d1361b2adb9 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 920.407313] env[61728]: DEBUG nova.compute.manager [req-0a2746f3-0be6-4cf5-9177-3eb8f68ebc9d req-f5e71ef9-e4b8-43af-bfaf-d8194ef85eee service nova] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Refreshing instance network info cache due to event network-changed-e06b1602-d9c2-459c-bdac-2d1361b2adb9. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 920.407577] env[61728]: DEBUG oslo_concurrency.lockutils [req-0a2746f3-0be6-4cf5-9177-3eb8f68ebc9d req-f5e71ef9-e4b8-43af-bfaf-d8194ef85eee service nova] Acquiring lock "refresh_cache-0e77ef23-3d1a-4f00-83f9-5b617334f980" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.407794] env[61728]: DEBUG oslo_concurrency.lockutils [req-0a2746f3-0be6-4cf5-9177-3eb8f68ebc9d req-f5e71ef9-e4b8-43af-bfaf-d8194ef85eee service nova] Acquired lock "refresh_cache-0e77ef23-3d1a-4f00-83f9-5b617334f980" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.408044] env[61728]: DEBUG nova.network.neutron [req-0a2746f3-0be6-4cf5-9177-3eb8f68ebc9d req-f5e71ef9-e4b8-43af-bfaf-d8194ef85eee service nova] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Refreshing network info cache for port e06b1602-d9c2-459c-bdac-2d1361b2adb9 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 920.620607] env[61728]: DEBUG nova.objects.instance [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lazy-loading 'numa_topology' on Instance uuid 3bfefcfc-db97-4a9d-86cb-9fb1d8158863 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.643688] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464483, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.689154] env[61728]: DEBUG oslo_concurrency.lockutils [None req-22efff5d-2b33-4932-b22a-6009da772bfe tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "40558dd0-69d3-42d6-87cf-8e4ae4ae9dae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.479s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.705545] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464486, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.915427] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f16fb5d1-241f-441e-b664-af1abfce7e15 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "ae25dc25-f600-471f-aaaa-f166ed691bf3" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.915708] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f16fb5d1-241f-441e-b664-af1abfce7e15 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "ae25dc25-f600-471f-aaaa-f166ed691bf3" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.915945] env[61728]: DEBUG nova.compute.manager [None req-f16fb5d1-241f-441e-b664-af1abfce7e15 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.916869] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eabec9eb-cd88-4651-ae59-30619506d784 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.925112] env[61728]: DEBUG nova.compute.manager [None req-f16fb5d1-241f-441e-b664-af1abfce7e15 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61728) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 920.925745] env[61728]: DEBUG nova.objects.instance [None req-f16fb5d1-241f-441e-b664-af1abfce7e15 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lazy-loading 'flavor' on Instance uuid ae25dc25-f600-471f-aaaa-f166ed691bf3 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.123521] env[61728]: DEBUG nova.objects.base [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Object Instance<3bfefcfc-db97-4a9d-86cb-9fb1d8158863> lazy-loaded attributes: resources,numa_topology {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 921.145908] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464483, 'name': ReconfigVM_Task, 'duration_secs': 0.779146} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.146059] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 2400abb1-8a07-4f6b-8818-778a2105beb8/2400abb1-8a07-4f6b-8818-778a2105beb8.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 921.149517] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bfc5d86b-8951-4030-89f0-00525ce5475c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.156894] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Waiting for the task: (returnval){ [ 921.156894] env[61728]: value = "task-464487" [ 921.156894] env[61728]: _type = "Task" [ 921.156894] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.169587] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464487, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.205296] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464486, 'name': CreateVM_Task, 'duration_secs': 0.709786} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.205760] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 921.206179] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.206351] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.206783] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 921.206938] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee06fa6e-a820-4457-a296-c1107d22acd7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.214366] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 921.214366] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5204536f-05d8-43b1-bcb8-1db56720ae0e" [ 921.214366] env[61728]: _type = "Task" [ 921.214366] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.226517] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5204536f-05d8-43b1-bcb8-1db56720ae0e, 'name': SearchDatastore_Task, 'duration_secs': 0.008861} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.227990] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.227990] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 921.227990] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.227990] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.227990] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.227990] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c3e755aa-60cd-4ae5-8481-e798a1b7a6fb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.235984] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.236202] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 921.239709] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ac52e81-ca4f-4d5e-9402-db76ea015345 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.245204] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 921.245204] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5228cef1-59ed-cd36-74c7-1e8b41650555" [ 921.245204] env[61728]: _type = "Task" [ 921.245204] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.257435] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5228cef1-59ed-cd36-74c7-1e8b41650555, 'name': SearchDatastore_Task, 'duration_secs': 0.007955} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.261767] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0ed207d-e9dd-42d7-993f-eeb225010e1d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.267556] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 921.267556] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527bcd61-9904-c578-4abc-1a09c336c44d" [ 921.267556] env[61728]: _type = "Task" [ 921.267556] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.277144] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527bcd61-9904-c578-4abc-1a09c336c44d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.327351] env[61728]: DEBUG nova.network.neutron [req-0a2746f3-0be6-4cf5-9177-3eb8f68ebc9d req-f5e71ef9-e4b8-43af-bfaf-d8194ef85eee service nova] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Updated VIF entry in instance network info cache for port e06b1602-d9c2-459c-bdac-2d1361b2adb9. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 921.327994] env[61728]: DEBUG nova.network.neutron [req-0a2746f3-0be6-4cf5-9177-3eb8f68ebc9d req-f5e71ef9-e4b8-43af-bfaf-d8194ef85eee service nova] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Updating instance_info_cache with network_info: [{"id": "e06b1602-d9c2-459c-bdac-2d1361b2adb9", "address": "fa:16:3e:60:91:58", "network": {"id": "02d8bb39-8b1a-4b90-8d0c-2873b2283fb5", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1139357359-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a75df31158574e6182e8d7505c659f1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16e15a36-a55b-4c27-b864-f284339009d0", "external-id": "nsx-vlan-transportzone-616", "segmentation_id": 616, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape06b1602-d9", "ovs_interfaceid": "e06b1602-d9c2-459c-bdac-2d1361b2adb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.433194] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f16fb5d1-241f-441e-b664-af1abfce7e15 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 921.434020] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca2fdb79-a295-4587-80e8-a795c130b50e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.445496] env[61728]: DEBUG oslo_vmware.api [None req-f16fb5d1-241f-441e-b664-af1abfce7e15 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 921.445496] env[61728]: value = "task-464488" [ 921.445496] env[61728]: _type = "Task" [ 921.445496] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.459402] env[61728]: DEBUG oslo_vmware.api [None req-f16fb5d1-241f-441e-b664-af1abfce7e15 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464488, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.484818] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6601be9d-d991-4db5-a1e7-8b5f259ee367 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.492875] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eae9c3a-b498-45a6-ac8c-d95182cf6ca2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.525328] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ce617c-df7e-4402-8885-7a08d59bde63 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.533656] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be3841cf-48b3-49ff-a421-fff5cb89d4ce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.549785] env[61728]: DEBUG nova.compute.provider_tree [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.667201] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464487, 'name': Rename_Task, 'duration_secs': 0.197795} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.667598] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 921.667898] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3e0ed2ec-be9d-4f66-a5fa-89edda34aed9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.674489] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Waiting for the task: (returnval){ [ 921.674489] env[61728]: value = "task-464489" [ 921.674489] env[61728]: _type = "Task" [ 921.674489] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.683725] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464489, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.780102] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527bcd61-9904-c578-4abc-1a09c336c44d, 'name': SearchDatastore_Task, 'duration_secs': 0.013792} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.780410] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.780697] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 0e77ef23-3d1a-4f00-83f9-5b617334f980/0e77ef23-3d1a-4f00-83f9-5b617334f980.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 921.780988] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a3084722-4c34-40a4-a497-219760440486 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.787297] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 921.787297] env[61728]: value = "task-464490" [ 921.787297] env[61728]: _type = "Task" [ 921.787297] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.797423] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464490, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.833954] env[61728]: DEBUG oslo_concurrency.lockutils [req-0a2746f3-0be6-4cf5-9177-3eb8f68ebc9d req-f5e71ef9-e4b8-43af-bfaf-d8194ef85eee service nova] Releasing lock "refresh_cache-0e77ef23-3d1a-4f00-83f9-5b617334f980" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.955568] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Successfully updated port: bcd321c5-6557-4c39-a5e4-025a4b20d7db {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 921.960165] env[61728]: DEBUG oslo_vmware.api [None req-f16fb5d1-241f-441e-b664-af1abfce7e15 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464488, 'name': PowerOffVM_Task, 'duration_secs': 0.322045} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.960320] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f16fb5d1-241f-441e-b664-af1abfce7e15 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 921.960563] env[61728]: DEBUG nova.compute.manager [None req-f16fb5d1-241f-441e-b664-af1abfce7e15 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.962013] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b448f3a5-923d-40e8-a691-57da170690fc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.055424] env[61728]: DEBUG nova.scheduler.client.report [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.185246] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464489, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.298123] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464490, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450111} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.298416] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 0e77ef23-3d1a-4f00-83f9-5b617334f980/0e77ef23-3d1a-4f00-83f9-5b617334f980.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 922.298660] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.298938] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-34bd7e41-3696-4329-a85d-fcd0d05d68ea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.305144] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 922.305144] env[61728]: value = "task-464491" [ 922.305144] env[61728]: _type = "Task" [ 922.305144] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.312916] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464491, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.433475] env[61728]: DEBUG nova.compute.manager [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Received event network-changed-51b6c967-37bc-4156-a23f-91040bb1c4a2 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 922.433702] env[61728]: DEBUG nova.compute.manager [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Refreshing instance network info cache due to event network-changed-51b6c967-37bc-4156-a23f-91040bb1c4a2. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 922.433927] env[61728]: DEBUG oslo_concurrency.lockutils [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] Acquiring lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.434096] env[61728]: DEBUG oslo_concurrency.lockutils [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] Acquired lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.434272] env[61728]: DEBUG nova.network.neutron [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Refreshing network info cache for port 51b6c967-37bc-4156-a23f-91040bb1c4a2 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 922.460973] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "refresh_cache-2c166dd3-985c-4637-92ab-939b46a7c90e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.461129] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquired lock "refresh_cache-2c166dd3-985c-4637-92ab-939b46a7c90e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.461280] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 922.477489] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f16fb5d1-241f-441e-b664-af1abfce7e15 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "ae25dc25-f600-471f-aaaa-f166ed691bf3" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.562s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.559953] env[61728]: DEBUG oslo_concurrency.lockutils [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.443s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.562432] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.287s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.564023] env[61728]: INFO nova.compute.claims [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 922.686958] env[61728]: DEBUG oslo_vmware.api [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464489, 'name': PowerOnVM_Task, 'duration_secs': 0.572444} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.687430] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 922.687684] env[61728]: INFO nova.compute.manager [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Took 9.56 seconds to spawn the instance on the hypervisor. [ 922.687917] env[61728]: DEBUG nova.compute.manager [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.688744] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c94a7ffa-1c3e-47ea-9431-ee7731be5c8f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.816504] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464491, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.126668} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.816806] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 922.817665] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d194dcbd-d86f-4244-9661-e536f650965f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.838939] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 0e77ef23-3d1a-4f00-83f9-5b617334f980/0e77ef23-3d1a-4f00-83f9-5b617334f980.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.839208] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c690dd6-2565-4c7c-8318-462b903fe3df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.858099] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 922.858099] env[61728]: value = "task-464492" [ 922.858099] env[61728]: _type = "Task" [ 922.858099] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.866384] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464492, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.989992] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 923.076977] env[61728]: DEBUG oslo_concurrency.lockutils [None req-11a6a522-e78b-4746-bc62-17541ab5081d tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 52.740s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.080544] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 30.014s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.080749] env[61728]: INFO nova.compute.manager [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Unshelving [ 923.123355] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Updating instance_info_cache with network_info: [{"id": "bcd321c5-6557-4c39-a5e4-025a4b20d7db", "address": "fa:16:3e:9d:c8:0f", "network": {"id": "02d8bb39-8b1a-4b90-8d0c-2873b2283fb5", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1139357359-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a75df31158574e6182e8d7505c659f1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16e15a36-a55b-4c27-b864-f284339009d0", "external-id": "nsx-vlan-transportzone-616", "segmentation_id": 616, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcd321c5-65", "ovs_interfaceid": "bcd321c5-6557-4c39-a5e4-025a4b20d7db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.207967] env[61728]: INFO nova.compute.manager [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Took 51.66 seconds to build instance. [ 923.247391] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.247805] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.273720] env[61728]: DEBUG nova.network.neutron [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Updated VIF entry in instance network info cache for port 51b6c967-37bc-4156-a23f-91040bb1c4a2. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 923.274164] env[61728]: DEBUG nova.network.neutron [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Updating instance_info_cache with network_info: [{"id": "51b6c967-37bc-4156-a23f-91040bb1c4a2", "address": "fa:16:3e:7f:f1:5d", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b6c967-37", "ovs_interfaceid": "51b6c967-37bc-4156-a23f-91040bb1c4a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.368233] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464492, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.583612] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "ae25dc25-f600-471f-aaaa-f166ed691bf3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.583954] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "ae25dc25-f600-471f-aaaa-f166ed691bf3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.584180] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "ae25dc25-f600-471f-aaaa-f166ed691bf3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.584391] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "ae25dc25-f600-471f-aaaa-f166ed691bf3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.584572] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "ae25dc25-f600-471f-aaaa-f166ed691bf3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.588506] env[61728]: INFO nova.compute.manager [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Terminating instance [ 923.592200] env[61728]: DEBUG nova.compute.manager [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 923.592382] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 923.594026] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693bbe6f-4b04-46dd-aa87-602fd32cde29 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.602235] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 923.602517] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-53da02eb-0908-409f-a975-6a2c57056edc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.625402] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Releasing lock "refresh_cache-2c166dd3-985c-4637-92ab-939b46a7c90e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.625730] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Instance network_info: |[{"id": "bcd321c5-6557-4c39-a5e4-025a4b20d7db", "address": "fa:16:3e:9d:c8:0f", "network": {"id": "02d8bb39-8b1a-4b90-8d0c-2873b2283fb5", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1139357359-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a75df31158574e6182e8d7505c659f1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16e15a36-a55b-4c27-b864-f284339009d0", "external-id": "nsx-vlan-transportzone-616", "segmentation_id": 616, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcd321c5-65", "ovs_interfaceid": "bcd321c5-6557-4c39-a5e4-025a4b20d7db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 923.626252] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:c8:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '16e15a36-a55b-4c27-b864-f284339009d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bcd321c5-6557-4c39-a5e4-025a4b20d7db', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 923.634640] env[61728]: DEBUG oslo.service.loopingcall [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.638379] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 923.640376] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6846ef8-adac-4cd9-89db-c5fa6dc3f4c7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.667928] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 923.667928] env[61728]: value = "task-464494" [ 923.667928] env[61728]: _type = "Task" [ 923.667928] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.676258] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 923.676421] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 923.676601] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleting the datastore file [datastore1] ae25dc25-f600-471f-aaaa-f166ed691bf3 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 923.676939] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f4932321-84ed-4f8f-a5a5-bb684fa2aac9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.683145] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464494, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.688553] env[61728]: DEBUG oslo_vmware.api [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 923.688553] env[61728]: value = "task-464495" [ 923.688553] env[61728]: _type = "Task" [ 923.688553] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.700210] env[61728]: DEBUG oslo_vmware.api [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464495, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.708288] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9d30ddba-918e-424f-b9b1-441c1b4e9d5c tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Lock "2400abb1-8a07-4f6b-8818-778a2105beb8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.514s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.750418] env[61728]: DEBUG nova.compute.manager [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 923.777243] env[61728]: DEBUG oslo_concurrency.lockutils [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] Releasing lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.777941] env[61728]: DEBUG nova.compute.manager [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Received event network-vif-plugged-bcd321c5-6557-4c39-a5e4-025a4b20d7db {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 923.778873] env[61728]: DEBUG oslo_concurrency.lockutils [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] Acquiring lock "2c166dd3-985c-4637-92ab-939b46a7c90e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.778873] env[61728]: DEBUG oslo_concurrency.lockutils [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] Lock "2c166dd3-985c-4637-92ab-939b46a7c90e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.778873] env[61728]: DEBUG oslo_concurrency.lockutils [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] Lock "2c166dd3-985c-4637-92ab-939b46a7c90e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.779121] env[61728]: DEBUG nova.compute.manager [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] No waiting events found dispatching network-vif-plugged-bcd321c5-6557-4c39-a5e4-025a4b20d7db {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 923.779250] env[61728]: WARNING nova.compute.manager [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Received unexpected event network-vif-plugged-bcd321c5-6557-4c39-a5e4-025a4b20d7db for instance with vm_state building and task_state spawning. [ 923.779480] env[61728]: DEBUG nova.compute.manager [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Received event network-changed-bcd321c5-6557-4c39-a5e4-025a4b20d7db {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 923.779885] env[61728]: DEBUG nova.compute.manager [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Refreshing instance network info cache due to event network-changed-bcd321c5-6557-4c39-a5e4-025a4b20d7db. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 923.780173] env[61728]: DEBUG oslo_concurrency.lockutils [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] Acquiring lock "refresh_cache-2c166dd3-985c-4637-92ab-939b46a7c90e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.780375] env[61728]: DEBUG oslo_concurrency.lockutils [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] Acquired lock "refresh_cache-2c166dd3-985c-4637-92ab-939b46a7c90e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.780606] env[61728]: DEBUG nova.network.neutron [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Refreshing network info cache for port bcd321c5-6557-4c39-a5e4-025a4b20d7db {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 923.871814] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464492, 'name': ReconfigVM_Task, 'duration_secs': 0.966336} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.872230] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 0e77ef23-3d1a-4f00-83f9-5b617334f980/0e77ef23-3d1a-4f00-83f9-5b617334f980.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.872960] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a5664ba-d95c-4e55-b16e-86d84b06d52f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.879401] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 923.879401] env[61728]: value = "task-464496" [ 923.879401] env[61728]: _type = "Task" [ 923.879401] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.890300] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464496, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.970233] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b8e5a5c-0f90-44dc-badb-6dd9803ef576 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.977928] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cbdd684-c3a5-4f0c-a200-74a7ae8bb83e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.008904] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e8c00c-74dd-43b6-8bd3-53f5b86efdb2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.019482] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e7cc4cf-7124-403f-8091-adc93c4df6eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.040884] env[61728]: DEBUG nova.compute.provider_tree [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.110270] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.178974] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464494, 'name': CreateVM_Task, 'duration_secs': 0.389424} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.179496] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 924.180259] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.180435] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.180825] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 924.181052] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f8d8346-b7d0-4b7a-bd78-259058f96655 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.185669] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 924.185669] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5217694a-fa89-b541-4cdb-b4ee64c94bb2" [ 924.185669] env[61728]: _type = "Task" [ 924.185669] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.193366] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5217694a-fa89-b541-4cdb-b4ee64c94bb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.200421] env[61728]: DEBUG oslo_vmware.api [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464495, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.207953} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.200663] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 924.200853] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 924.201078] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 924.201260] env[61728]: INFO nova.compute.manager [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Took 0.61 seconds to destroy the instance on the hypervisor. [ 924.201494] env[61728]: DEBUG oslo.service.loopingcall [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.201709] env[61728]: DEBUG nova.compute.manager [-] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 924.201831] env[61728]: DEBUG nova.network.neutron [-] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 924.271513] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.394763] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464496, 'name': Rename_Task, 'duration_secs': 0.154717} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.395263] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 924.395608] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-86e3a68e-80a1-46f0-bbb7-6357db54186c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.403110] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 924.403110] env[61728]: value = "task-464497" [ 924.403110] env[61728]: _type = "Task" [ 924.403110] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.411558] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464497, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.490722] env[61728]: DEBUG nova.network.neutron [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Updated VIF entry in instance network info cache for port bcd321c5-6557-4c39-a5e4-025a4b20d7db. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 924.490722] env[61728]: DEBUG nova.network.neutron [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Updating instance_info_cache with network_info: [{"id": "bcd321c5-6557-4c39-a5e4-025a4b20d7db", "address": "fa:16:3e:9d:c8:0f", "network": {"id": "02d8bb39-8b1a-4b90-8d0c-2873b2283fb5", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1139357359-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a75df31158574e6182e8d7505c659f1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16e15a36-a55b-4c27-b864-f284339009d0", "external-id": "nsx-vlan-transportzone-616", "segmentation_id": 616, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcd321c5-65", "ovs_interfaceid": "bcd321c5-6557-4c39-a5e4-025a4b20d7db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.543639] env[61728]: DEBUG nova.scheduler.client.report [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 924.574820] env[61728]: DEBUG nova.compute.manager [req-3a4ac1f0-6d25-4ffa-a9bd-65ce7e872a34 req-4292fa35-70eb-4ce7-b2a8-36d8b8d46151 service nova] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Received event network-vif-deleted-54951cdd-ce11-45d8-9791-d96e06eb10f6 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 924.575105] env[61728]: INFO nova.compute.manager [req-3a4ac1f0-6d25-4ffa-a9bd-65ce7e872a34 req-4292fa35-70eb-4ce7-b2a8-36d8b8d46151 service nova] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Neutron deleted interface 54951cdd-ce11-45d8-9791-d96e06eb10f6; detaching it from the instance and deleting it from the info cache [ 924.575326] env[61728]: DEBUG nova.network.neutron [req-3a4ac1f0-6d25-4ffa-a9bd-65ce7e872a34 req-4292fa35-70eb-4ce7-b2a8-36d8b8d46151 service nova] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.697147] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5217694a-fa89-b541-4cdb-b4ee64c94bb2, 'name': SearchDatastore_Task, 'duration_secs': 0.00923} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.697595] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.697647] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 924.697984] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.698278] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.698541] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.698878] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9950bb4c-f763-468a-9a3b-d7a0a74d1fed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.707521] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.707964] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 924.708501] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dba16d1f-c7f3-4989-a7d9-646f6a9dc52c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.714202] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 924.714202] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ace19f-70fa-bccf-42c0-822df254968b" [ 924.714202] env[61728]: _type = "Task" [ 924.714202] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.722126] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ace19f-70fa-bccf-42c0-822df254968b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.920158] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464497, 'name': PowerOnVM_Task, 'duration_secs': 0.504556} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.922880] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 924.924383] env[61728]: INFO nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Took 7.40 seconds to spawn the instance on the hypervisor. [ 924.925594] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.929925] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8f5e25-b12c-48dd-87ca-e77f758e1d70 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.995961] env[61728]: DEBUG oslo_concurrency.lockutils [req-e797c674-db08-4dec-b6a6-1af871e78cc6 req-62d5c2ad-e893-4864-95ba-d8ab3c1191ca service nova] Releasing lock "refresh_cache-2c166dd3-985c-4637-92ab-939b46a7c90e" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.028926] env[61728]: DEBUG nova.network.neutron [-] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.048757] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.049687] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 925.052615] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.592s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.053847] env[61728]: DEBUG nova.objects.instance [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lazy-loading 'resources' on Instance uuid 121bef05-ef7b-47ee-b737-15f8b9d91158 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.078896] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e59fbe9-c7aa-4f8d-b8c5-b75bb0d33738 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.090031] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec322101-4aec-4fc9-8144-45a40407e643 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.119153] env[61728]: DEBUG nova.compute.manager [req-3a4ac1f0-6d25-4ffa-a9bd-65ce7e872a34 req-4292fa35-70eb-4ce7-b2a8-36d8b8d46151 service nova] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Detach interface failed, port_id=54951cdd-ce11-45d8-9791-d96e06eb10f6, reason: Instance ae25dc25-f600-471f-aaaa-f166ed691bf3 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 925.224321] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ace19f-70fa-bccf-42c0-822df254968b, 'name': SearchDatastore_Task, 'duration_secs': 0.008702} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.225105] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-178f6bc9-17e9-4c39-bdb3-52272f44184a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.230021] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 925.230021] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a77f30-8f2a-7d77-3104-ffa7d366cb95" [ 925.230021] env[61728]: _type = "Task" [ 925.230021] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.237704] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a77f30-8f2a-7d77-3104-ffa7d366cb95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.284493] env[61728]: DEBUG oslo_concurrency.lockutils [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Acquiring lock "2400abb1-8a07-4f6b-8818-778a2105beb8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.284831] env[61728]: DEBUG oslo_concurrency.lockutils [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Lock "2400abb1-8a07-4f6b-8818-778a2105beb8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.285118] env[61728]: DEBUG oslo_concurrency.lockutils [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Acquiring lock "2400abb1-8a07-4f6b-8818-778a2105beb8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.285353] env[61728]: DEBUG oslo_concurrency.lockutils [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Lock "2400abb1-8a07-4f6b-8818-778a2105beb8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.285556] env[61728]: DEBUG oslo_concurrency.lockutils [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Lock "2400abb1-8a07-4f6b-8818-778a2105beb8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.287970] env[61728]: INFO nova.compute.manager [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Terminating instance [ 925.289850] env[61728]: DEBUG nova.compute.manager [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 925.290082] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 925.290912] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e3492b5-3bf7-4625-a2d0-d2f4792babc9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.297964] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 925.298196] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3bd7d78a-54af-4b63-be66-9d6ae14099ac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.304126] env[61728]: DEBUG oslo_vmware.api [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Waiting for the task: (returnval){ [ 925.304126] env[61728]: value = "task-464498" [ 925.304126] env[61728]: _type = "Task" [ 925.304126] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.311580] env[61728]: DEBUG oslo_vmware.api [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464498, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.474321] env[61728]: INFO nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Took 41.18 seconds to build instance. [ 925.531557] env[61728]: INFO nova.compute.manager [-] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Took 1.33 seconds to deallocate network for instance. [ 925.557116] env[61728]: DEBUG nova.compute.utils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 925.560682] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 925.561060] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 925.649926] env[61728]: DEBUG nova.policy [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'df4f3ee920b04018acf4c3c70aca7e55', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a75df31158574e6182e8d7505c659f1c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 925.743062] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a77f30-8f2a-7d77-3104-ffa7d366cb95, 'name': SearchDatastore_Task, 'duration_secs': 0.012127} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.743468] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.743976] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2c166dd3-985c-4637-92ab-939b46a7c90e/2c166dd3-985c-4637-92ab-939b46a7c90e.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 925.744353] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-33cd34cb-3222-43bd-8dbf-79780dbbdecd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.756643] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 925.756643] env[61728]: value = "task-464499" [ 925.756643] env[61728]: _type = "Task" [ 925.756643] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.767549] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464499, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.816218] env[61728]: DEBUG oslo_vmware.api [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464498, 'name': PowerOffVM_Task, 'duration_secs': 0.502611} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.817261] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 925.817483] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 925.817761] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-765e30b4-2927-47e1-9e96-8873c716f2e3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.878966] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 925.879046] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 925.879209] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Deleting the datastore file [datastore1] 2400abb1-8a07-4f6b-8818-778a2105beb8 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.879497] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-494665de-02c6-4056-b3df-2352a486c881 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.886511] env[61728]: DEBUG oslo_vmware.api [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Waiting for the task: (returnval){ [ 925.886511] env[61728]: value = "task-464501" [ 925.886511] env[61728]: _type = "Task" [ 925.886511] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.897265] env[61728]: DEBUG oslo_vmware.api [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464501, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.903136] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc78a99-fd11-43ca-aa6c-e2894bd7a132 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.909863] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df95bf2-5248-42fd-b29c-eeedddb3e70d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.940776] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5408ea-7b10-41d9-8562-fa8d7dd50ad0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.952029] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39442a79-fff7-4819-8977-7b9ae0215883 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.968991] env[61728]: DEBUG nova.compute.provider_tree [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.973891] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Successfully created port: 05dc7f10-50f6-4a11-95b8-a6bbd35821b5 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 925.976948] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "0e77ef23-3d1a-4f00-83f9-5b617334f980" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.134s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.038962] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.061065] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 926.268334] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464499, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.396314] env[61728]: DEBUG oslo_vmware.api [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Task: {'id': task-464501, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.462185} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.396624] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.396820] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 926.397015] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 926.397206] env[61728]: INFO nova.compute.manager [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Took 1.11 seconds to destroy the instance on the hypervisor. [ 926.397459] env[61728]: DEBUG oslo.service.loopingcall [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.397663] env[61728]: DEBUG nova.compute.manager [-] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 926.397759] env[61728]: DEBUG nova.network.neutron [-] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 926.479336] env[61728]: DEBUG nova.scheduler.client.report [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 926.738970] env[61728]: DEBUG nova.compute.manager [req-0ca6742c-39f2-43b4-bfa6-618afc9b144a req-84a07de7-12ca-4f91-a8e1-e8110976ef02 service nova] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Received event network-vif-deleted-dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 926.739205] env[61728]: INFO nova.compute.manager [req-0ca6742c-39f2-43b4-bfa6-618afc9b144a req-84a07de7-12ca-4f91-a8e1-e8110976ef02 service nova] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Neutron deleted interface dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9; detaching it from the instance and deleting it from the info cache [ 926.739348] env[61728]: DEBUG nova.network.neutron [req-0ca6742c-39f2-43b4-bfa6-618afc9b144a req-84a07de7-12ca-4f91-a8e1-e8110976ef02 service nova] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.767376] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464499, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.590394} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.767646] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2c166dd3-985c-4637-92ab-939b46a7c90e/2c166dd3-985c-4637-92ab-939b46a7c90e.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 926.767952] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 926.768123] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bf9e3155-f188-4b09-b1de-90083ebee68b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.776956] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 926.776956] env[61728]: value = "task-464502" [ 926.776956] env[61728]: _type = "Task" [ 926.776956] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.783046] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464502, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.980856] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.928s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.983255] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.286s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.984087] env[61728]: DEBUG nova.objects.instance [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Lazy-loading 'resources' on Instance uuid 174607a1-9bc8-4e07-8993-7f0bb0f308e9 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.004784] env[61728]: INFO nova.scheduler.client.report [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Deleted allocations for instance 121bef05-ef7b-47ee-b737-15f8b9d91158 [ 927.071687] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 927.096073] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 927.096360] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 927.096524] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.096712] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 927.096866] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.097036] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 927.097277] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 927.097447] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 927.097623] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 927.097790] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 927.097967] env[61728]: DEBUG nova.virt.hardware [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 927.099127] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a57fe08-fa0c-4206-bf08-3f6ed6e73920 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.107129] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ce0676-5835-40a4-93c6-8319d92c002c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.125656] env[61728]: DEBUG nova.network.neutron [-] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.242428] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7db51698-e6d1-4b59-95eb-35b2573547b4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.252943] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1243a7-744e-4714-beb6-1f923170ef17 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.283502] env[61728]: DEBUG nova.compute.manager [req-0ca6742c-39f2-43b4-bfa6-618afc9b144a req-84a07de7-12ca-4f91-a8e1-e8110976ef02 service nova] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Detach interface failed, port_id=dd1de7cf-9c17-4bfd-a07a-3278b3f1d9f9, reason: Instance 2400abb1-8a07-4f6b-8818-778a2105beb8 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 927.292046] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464502, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067046} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.292425] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 927.293255] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4bc0d83-9472-4e7b-996d-ed3fd91ba094 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.316879] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 2c166dd3-985c-4637-92ab-939b46a7c90e/2c166dd3-985c-4637-92ab-939b46a7c90e.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.317202] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0c161a2-961a-4d2b-af42-879580418e86 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.336434] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 927.336434] env[61728]: value = "task-464503" [ 927.336434] env[61728]: _type = "Task" [ 927.336434] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.344252] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464503, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.513840] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f97f0f32-e221-4b9e-81e8-d7dc5e4d6142 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "121bef05-ef7b-47ee-b737-15f8b9d91158" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.487s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.628533] env[61728]: INFO nova.compute.manager [-] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Took 1.23 seconds to deallocate network for instance. [ 927.797164] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac264ff-ec32-4752-afad-8d84b539eabe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.809253] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d491e577-9e52-476b-9fed-6b63dd67117f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.854447] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0574aa4-6d75-4fb1-9f50-a0284ede4f00 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.862321] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464503, 'name': ReconfigVM_Task, 'duration_secs': 0.262315} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.864424] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 2c166dd3-985c-4637-92ab-939b46a7c90e/2c166dd3-985c-4637-92ab-939b46a7c90e.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.865091] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dff427dd-8829-45ec-9b7b-42c976e8ddfd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.867459] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ddf9fc-e532-4300-8c75-7c915840b376 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.883851] env[61728]: DEBUG nova.compute.provider_tree [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.886427] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 927.886427] env[61728]: value = "task-464504" [ 927.886427] env[61728]: _type = "Task" [ 927.886427] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.895805] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464504, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.066148] env[61728]: DEBUG nova.compute.manager [req-740c5101-b4fa-4cab-ae4a-79b07b277b94 req-b559d699-b2e3-4c7c-a1d6-79943c91b298 service nova] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Received event network-vif-plugged-05dc7f10-50f6-4a11-95b8-a6bbd35821b5 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 928.066148] env[61728]: DEBUG oslo_concurrency.lockutils [req-740c5101-b4fa-4cab-ae4a-79b07b277b94 req-b559d699-b2e3-4c7c-a1d6-79943c91b298 service nova] Acquiring lock "ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.066148] env[61728]: DEBUG oslo_concurrency.lockutils [req-740c5101-b4fa-4cab-ae4a-79b07b277b94 req-b559d699-b2e3-4c7c-a1d6-79943c91b298 service nova] Lock "ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.066148] env[61728]: DEBUG oslo_concurrency.lockutils [req-740c5101-b4fa-4cab-ae4a-79b07b277b94 req-b559d699-b2e3-4c7c-a1d6-79943c91b298 service nova] Lock "ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.066148] env[61728]: DEBUG nova.compute.manager [req-740c5101-b4fa-4cab-ae4a-79b07b277b94 req-b559d699-b2e3-4c7c-a1d6-79943c91b298 service nova] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] No waiting events found dispatching network-vif-plugged-05dc7f10-50f6-4a11-95b8-a6bbd35821b5 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 928.066148] env[61728]: WARNING nova.compute.manager [req-740c5101-b4fa-4cab-ae4a-79b07b277b94 req-b559d699-b2e3-4c7c-a1d6-79943c91b298 service nova] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Received unexpected event network-vif-plugged-05dc7f10-50f6-4a11-95b8-a6bbd35821b5 for instance with vm_state building and task_state spawning. [ 928.140572] env[61728]: DEBUG oslo_concurrency.lockutils [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.152448] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Successfully updated port: 05dc7f10-50f6-4a11-95b8-a6bbd35821b5 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 928.395050] env[61728]: DEBUG nova.scheduler.client.report [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.400903] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464504, 'name': Rename_Task, 'duration_secs': 0.157433} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.401438] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 928.401787] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05d5cfff-0d14-4c0e-8d2c-9424157f1203 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.408289] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 928.408289] env[61728]: value = "task-464505" [ 928.408289] env[61728]: _type = "Task" [ 928.408289] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.418406] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464505, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.659042] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "refresh_cache-ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.659170] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquired lock "refresh_cache-ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.659335] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 928.898038] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.914s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.900341] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.464s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.901939] env[61728]: INFO nova.compute.claims [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 928.920021] env[61728]: INFO nova.scheduler.client.report [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Deleted allocations for instance 174607a1-9bc8-4e07-8993-7f0bb0f308e9 [ 928.923868] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464505, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.139577] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "db547fea-4d44-440e-ac3e-29ae9b2c547b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.139840] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "db547fea-4d44-440e-ac3e-29ae9b2c547b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.171276] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "0ff847f9-97d0-40d9-a458-038fe36b5c62" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.171639] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "0ff847f9-97d0-40d9-a458-038fe36b5c62" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.197290] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 929.401728] env[61728]: DEBUG nova.network.neutron [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Updating instance_info_cache with network_info: [{"id": "05dc7f10-50f6-4a11-95b8-a6bbd35821b5", "address": "fa:16:3e:12:c6:73", "network": {"id": "02d8bb39-8b1a-4b90-8d0c-2873b2283fb5", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1139357359-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a75df31158574e6182e8d7505c659f1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16e15a36-a55b-4c27-b864-f284339009d0", "external-id": "nsx-vlan-transportzone-616", "segmentation_id": 616, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05dc7f10-50", "ovs_interfaceid": "05dc7f10-50f6-4a11-95b8-a6bbd35821b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.423385] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464505, 'name': PowerOnVM_Task, 'duration_secs': 0.602439} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.423669] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 929.423912] env[61728]: INFO nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Took 9.33 seconds to spawn the instance on the hypervisor. [ 929.424143] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.424933] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7de861-28c5-4b14-87fb-11008faa973a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.436025] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58c94790-c42b-40f8-86a6-0fc4ced737c6 tempest-ServersTestManualDisk-1692275760 tempest-ServersTestManualDisk-1692275760-project-member] Lock "174607a1-9bc8-4e07-8993-7f0bb0f308e9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 43.148s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.642181] env[61728]: DEBUG nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 929.673828] env[61728]: DEBUG nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 929.904347] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Releasing lock "refresh_cache-ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.904842] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Instance network_info: |[{"id": "05dc7f10-50f6-4a11-95b8-a6bbd35821b5", "address": "fa:16:3e:12:c6:73", "network": {"id": "02d8bb39-8b1a-4b90-8d0c-2873b2283fb5", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1139357359-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a75df31158574e6182e8d7505c659f1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16e15a36-a55b-4c27-b864-f284339009d0", "external-id": "nsx-vlan-transportzone-616", "segmentation_id": 616, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05dc7f10-50", "ovs_interfaceid": "05dc7f10-50f6-4a11-95b8-a6bbd35821b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 929.905168] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:12:c6:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '16e15a36-a55b-4c27-b864-f284339009d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '05dc7f10-50f6-4a11-95b8-a6bbd35821b5', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 929.912690] env[61728]: DEBUG oslo.service.loopingcall [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.915910] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 929.916389] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ef6ae07c-07c4-4697-b623-c01f058c1072 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.949740] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 929.949740] env[61728]: value = "task-464506" [ 929.949740] env[61728]: _type = "Task" [ 929.949740] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.950717] env[61728]: INFO nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Took 42.92 seconds to build instance. [ 929.962050] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464506, 'name': CreateVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.162042] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.193827] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.258145] env[61728]: DEBUG nova.compute.manager [req-11bc9f6e-f4dd-49ca-8263-10a13d24c516 req-0b4a2a58-c3f6-4c7a-921b-a17caecd619a service nova] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Received event network-changed-05dc7f10-50f6-4a11-95b8-a6bbd35821b5 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 930.258145] env[61728]: DEBUG nova.compute.manager [req-11bc9f6e-f4dd-49ca-8263-10a13d24c516 req-0b4a2a58-c3f6-4c7a-921b-a17caecd619a service nova] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Refreshing instance network info cache due to event network-changed-05dc7f10-50f6-4a11-95b8-a6bbd35821b5. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 930.258145] env[61728]: DEBUG oslo_concurrency.lockutils [req-11bc9f6e-f4dd-49ca-8263-10a13d24c516 req-0b4a2a58-c3f6-4c7a-921b-a17caecd619a service nova] Acquiring lock "refresh_cache-ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.258259] env[61728]: DEBUG oslo_concurrency.lockutils [req-11bc9f6e-f4dd-49ca-8263-10a13d24c516 req-0b4a2a58-c3f6-4c7a-921b-a17caecd619a service nova] Acquired lock "refresh_cache-ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.258432] env[61728]: DEBUG nova.network.neutron [req-11bc9f6e-f4dd-49ca-8263-10a13d24c516 req-0b4a2a58-c3f6-4c7a-921b-a17caecd619a service nova] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Refreshing network info cache for port 05dc7f10-50f6-4a11-95b8-a6bbd35821b5 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 930.391570] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae66e204-ca7d-4801-8e8a-3ad21c8a40ee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.399165] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2635c774-7182-4cf3-b8e0-a031a1dac0f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.434882] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f39ba02f-1294-477b-a1f3-453d666ba5cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.442911] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e7b151-8eff-4ecd-b39c-4d44eff6364b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.458408] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "2c166dd3-985c-4637-92ab-939b46a7c90e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.581s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.458408] env[61728]: DEBUG nova.compute.provider_tree [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.466192] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464506, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.962584] env[61728]: DEBUG nova.scheduler.client.report [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.969562] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464506, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.086578] env[61728]: DEBUG nova.network.neutron [req-11bc9f6e-f4dd-49ca-8263-10a13d24c516 req-0b4a2a58-c3f6-4c7a-921b-a17caecd619a service nova] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Updated VIF entry in instance network info cache for port 05dc7f10-50f6-4a11-95b8-a6bbd35821b5. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 931.086978] env[61728]: DEBUG nova.network.neutron [req-11bc9f6e-f4dd-49ca-8263-10a13d24c516 req-0b4a2a58-c3f6-4c7a-921b-a17caecd619a service nova] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Updating instance_info_cache with network_info: [{"id": "05dc7f10-50f6-4a11-95b8-a6bbd35821b5", "address": "fa:16:3e:12:c6:73", "network": {"id": "02d8bb39-8b1a-4b90-8d0c-2873b2283fb5", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1139357359-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a75df31158574e6182e8d7505c659f1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16e15a36-a55b-4c27-b864-f284339009d0", "external-id": "nsx-vlan-transportzone-616", "segmentation_id": 616, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05dc7f10-50", "ovs_interfaceid": "05dc7f10-50f6-4a11-95b8-a6bbd35821b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.467651] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464506, 'name': CreateVM_Task, 'duration_secs': 1.21866} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.467836] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 931.468554] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.468734] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.469082] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 931.469345] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c25987ff-8097-4c12-b911-85489f464dae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.471488] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.571s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.471984] env[61728]: DEBUG nova.compute.manager [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 931.474591] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.796s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.476022] env[61728]: INFO nova.compute.claims [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.484023] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 931.484023] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52706e0e-6f2a-2d88-ed8d-d02c40452ba1" [ 931.484023] env[61728]: _type = "Task" [ 931.484023] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.494374] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52706e0e-6f2a-2d88-ed8d-d02c40452ba1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.590661] env[61728]: DEBUG oslo_concurrency.lockutils [req-11bc9f6e-f4dd-49ca-8263-10a13d24c516 req-0b4a2a58-c3f6-4c7a-921b-a17caecd619a service nova] Releasing lock "refresh_cache-ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.984897] env[61728]: DEBUG nova.compute.utils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 931.986385] env[61728]: DEBUG nova.compute.manager [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 931.986561] env[61728]: DEBUG nova.network.neutron [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 931.999119] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52706e0e-6f2a-2d88-ed8d-d02c40452ba1, 'name': SearchDatastore_Task, 'duration_secs': 0.009799} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.999119] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.999119] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 931.999439] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.999528] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.999681] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 932.000171] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ef1a52f-01ec-4ebc-a472-b7e7e6ebaf36 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.008223] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 932.008540] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 932.009893] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2aec969-fb81-40b9-884d-485d4b02d254 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.016269] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 932.016269] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527393ea-e211-2490-fd16-18296818be82" [ 932.016269] env[61728]: _type = "Task" [ 932.016269] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.024019] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527393ea-e211-2490-fd16-18296818be82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.034673] env[61728]: DEBUG nova.policy [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67d7c6a6191045cf85f790acebaf811e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7f61c87df8b485b9a6370032dd7af37', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 932.312730] env[61728]: DEBUG nova.network.neutron [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Successfully created port: 4660e4da-30e8-4e51-9aea-a9375732537f {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 932.490340] env[61728]: DEBUG nova.compute.manager [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 932.529275] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527393ea-e211-2490-fd16-18296818be82, 'name': SearchDatastore_Task, 'duration_secs': 0.008593} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.530701] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4de332f5-bf41-470e-bca0-23d4d454b4c3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.543158] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 932.543158] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f0a460-535f-7938-bb3f-694465b6f0b0" [ 932.543158] env[61728]: _type = "Task" [ 932.543158] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.551055] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f0a460-535f-7938-bb3f-694465b6f0b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.880057] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1d5eba-d00b-417b-80c0-aaf12c0ca795 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.888425] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d51e09e0-fac2-4e8a-b82c-bec6239cf2e8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.923173] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a1683b-1af5-472d-9ded-ce4f51c42461 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.930744] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4bfdc6-1786-4115-8810-d2d7ba5e3e8a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.949524] env[61728]: DEBUG nova.compute.provider_tree [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.060604] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f0a460-535f-7938-bb3f-694465b6f0b0, 'name': SearchDatastore_Task, 'duration_secs': 0.039625} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.061516] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.062748] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69/ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 933.062748] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bee4164e-7cbf-4af5-9034-1cf6d3bd539c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.071937] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 933.071937] env[61728]: value = "task-464507" [ 933.071937] env[61728]: _type = "Task" [ 933.071937] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.082660] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464507, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.451532] env[61728]: DEBUG nova.scheduler.client.report [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.501439] env[61728]: DEBUG nova.compute.manager [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 933.536928] env[61728]: DEBUG nova.virt.hardware [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.537397] env[61728]: DEBUG nova.virt.hardware [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.537397] env[61728]: DEBUG nova.virt.hardware [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.537658] env[61728]: DEBUG nova.virt.hardware [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.537943] env[61728]: DEBUG nova.virt.hardware [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.538145] env[61728]: DEBUG nova.virt.hardware [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.538495] env[61728]: DEBUG nova.virt.hardware [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.538495] env[61728]: DEBUG nova.virt.hardware [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.538675] env[61728]: DEBUG nova.virt.hardware [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.539054] env[61728]: DEBUG nova.virt.hardware [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.539250] env[61728]: DEBUG nova.virt.hardware [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.540551] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a928ad-8c62-462a-acbf-051046c95c5f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.548974] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdbb22a5-2f9e-4af9-92d5-1d2a35006367 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.584193] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464507, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.956976] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.957608] env[61728]: DEBUG nova.compute.manager [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 933.964020] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 34.986s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.082889] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464507, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.857065} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.083275] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69/ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 934.083434] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 934.083556] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e008b51-d3b0-45b7-a681-ed7037484887 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.095782] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 934.095782] env[61728]: value = "task-464508" [ 934.095782] env[61728]: _type = "Task" [ 934.095782] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.107175] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464508, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.465211] env[61728]: DEBUG nova.compute.utils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.466663] env[61728]: DEBUG nova.compute.manager [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 934.466873] env[61728]: DEBUG nova.network.neutron [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 934.474199] env[61728]: INFO nova.compute.claims [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 934.520041] env[61728]: DEBUG nova.policy [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54c30aeeea3a4c1b8ed34430361c2a4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a02bd852c564acd99e12fd17279f028', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 934.591468] env[61728]: DEBUG nova.network.neutron [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Successfully updated port: 4660e4da-30e8-4e51-9aea-a9375732537f {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 934.616116] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464508, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067213} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.616489] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 934.617385] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fce8575-762e-4fe6-b317-375ac104df92 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.647488] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69/ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.647488] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47db1866-fa57-40f9-b88c-3bd7c1a5bd70 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.671983] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 934.671983] env[61728]: value = "task-464509" [ 934.671983] env[61728]: _type = "Task" [ 934.671983] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.679993] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464509, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.711224] env[61728]: DEBUG nova.compute.manager [req-09cfa3a3-bed0-43fb-a6f5-e4248316a2c2 req-18c55c5f-216a-4df7-920c-2079c9bbf48b service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Received event network-vif-plugged-4660e4da-30e8-4e51-9aea-a9375732537f {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 934.711468] env[61728]: DEBUG oslo_concurrency.lockutils [req-09cfa3a3-bed0-43fb-a6f5-e4248316a2c2 req-18c55c5f-216a-4df7-920c-2079c9bbf48b service nova] Acquiring lock "8c5b2380-bd14-451e-9612-2767c2a6adb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.712516] env[61728]: DEBUG oslo_concurrency.lockutils [req-09cfa3a3-bed0-43fb-a6f5-e4248316a2c2 req-18c55c5f-216a-4df7-920c-2079c9bbf48b service nova] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.716761] env[61728]: DEBUG oslo_concurrency.lockutils [req-09cfa3a3-bed0-43fb-a6f5-e4248316a2c2 req-18c55c5f-216a-4df7-920c-2079c9bbf48b service nova] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.716761] env[61728]: DEBUG nova.compute.manager [req-09cfa3a3-bed0-43fb-a6f5-e4248316a2c2 req-18c55c5f-216a-4df7-920c-2079c9bbf48b service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] No waiting events found dispatching network-vif-plugged-4660e4da-30e8-4e51-9aea-a9375732537f {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 934.716761] env[61728]: WARNING nova.compute.manager [req-09cfa3a3-bed0-43fb-a6f5-e4248316a2c2 req-18c55c5f-216a-4df7-920c-2079c9bbf48b service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Received unexpected event network-vif-plugged-4660e4da-30e8-4e51-9aea-a9375732537f for instance with vm_state building and task_state spawning. [ 934.802028] env[61728]: DEBUG nova.network.neutron [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Successfully created port: fe024a36-a6a9-44ed-b043-1195882a2ae7 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.879895] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Acquiring lock "7e8cde9c-890c-4f50-9ccb-46aa71220395" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.880256] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Lock "7e8cde9c-890c-4f50-9ccb-46aa71220395" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.971575] env[61728]: DEBUG nova.compute.manager [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 934.980806] env[61728]: INFO nova.compute.resource_tracker [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating resource usage from migration 92379aa6-c648-4aae-8141-f8deadfff31e [ 935.097408] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.097408] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.097408] env[61728]: DEBUG nova.network.neutron [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 935.185123] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464509, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.371550] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6cd5bc2-a6eb-4430-8b14-e103b13797f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.381045] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbfaff4b-00de-4d21-8de0-81635381fc8f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.384550] env[61728]: DEBUG nova.compute.manager [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 935.416400] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634c6a52-b4af-4e79-8491-7648dd96369e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.424882] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff74f837-03fd-45f0-9ea5-835d886a0ece {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.439641] env[61728]: DEBUG nova.compute.provider_tree [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.671557] env[61728]: DEBUG nova.network.neutron [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 935.684017] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464509, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.911752] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.945193] env[61728]: DEBUG nova.scheduler.client.report [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.950880] env[61728]: DEBUG nova.network.neutron [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance_info_cache with network_info: [{"id": "4660e4da-30e8-4e51-9aea-a9375732537f", "address": "fa:16:3e:e1:e9:0c", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4660e4da-30", "ovs_interfaceid": "4660e4da-30e8-4e51-9aea-a9375732537f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.983432] env[61728]: DEBUG nova.compute.manager [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 936.014810] env[61728]: DEBUG nova.virt.hardware [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.015094] env[61728]: DEBUG nova.virt.hardware [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.015265] env[61728]: DEBUG nova.virt.hardware [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.015457] env[61728]: DEBUG nova.virt.hardware [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.015614] env[61728]: DEBUG nova.virt.hardware [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.015765] env[61728]: DEBUG nova.virt.hardware [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.015988] env[61728]: DEBUG nova.virt.hardware [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.020746] env[61728]: DEBUG nova.virt.hardware [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.020970] env[61728]: DEBUG nova.virt.hardware [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.021208] env[61728]: DEBUG nova.virt.hardware [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.021346] env[61728]: DEBUG nova.virt.hardware [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.022255] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50646df3-0419-4a87-9ef6-9c235a570323 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.034266] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c3427a-d1c3-4d37-88be-3fd3fd3450ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.184204] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464509, 'name': ReconfigVM_Task, 'duration_secs': 1.150322} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.184523] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Reconfigured VM instance instance-00000057 to attach disk [datastore1] ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69/ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.185179] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9415de74-8d91-4bb5-b0f1-3bd0b0a1d1f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.192022] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 936.192022] env[61728]: value = "task-464510" [ 936.192022] env[61728]: _type = "Task" [ 936.192022] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.200820] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464510, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.273321] env[61728]: DEBUG nova.compute.manager [req-279a211d-afec-41dd-a47e-0bde708b91ef req-efa7b3b6-2805-4d72-ba2e-24ef059201a1 service nova] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Received event network-vif-plugged-fe024a36-a6a9-44ed-b043-1195882a2ae7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 936.273321] env[61728]: DEBUG oslo_concurrency.lockutils [req-279a211d-afec-41dd-a47e-0bde708b91ef req-efa7b3b6-2805-4d72-ba2e-24ef059201a1 service nova] Acquiring lock "aebc4c7c-09d6-484f-8ad6-2416b3a41828-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.273321] env[61728]: DEBUG oslo_concurrency.lockutils [req-279a211d-afec-41dd-a47e-0bde708b91ef req-efa7b3b6-2805-4d72-ba2e-24ef059201a1 service nova] Lock "aebc4c7c-09d6-484f-8ad6-2416b3a41828-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.274030] env[61728]: DEBUG oslo_concurrency.lockutils [req-279a211d-afec-41dd-a47e-0bde708b91ef req-efa7b3b6-2805-4d72-ba2e-24ef059201a1 service nova] Lock "aebc4c7c-09d6-484f-8ad6-2416b3a41828-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.274394] env[61728]: DEBUG nova.compute.manager [req-279a211d-afec-41dd-a47e-0bde708b91ef req-efa7b3b6-2805-4d72-ba2e-24ef059201a1 service nova] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] No waiting events found dispatching network-vif-plugged-fe024a36-a6a9-44ed-b043-1195882a2ae7 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 936.274662] env[61728]: WARNING nova.compute.manager [req-279a211d-afec-41dd-a47e-0bde708b91ef req-efa7b3b6-2805-4d72-ba2e-24ef059201a1 service nova] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Received unexpected event network-vif-plugged-fe024a36-a6a9-44ed-b043-1195882a2ae7 for instance with vm_state building and task_state spawning. [ 936.354548] env[61728]: DEBUG nova.network.neutron [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Successfully updated port: fe024a36-a6a9-44ed-b043-1195882a2ae7 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 936.447975] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.487s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.448427] env[61728]: INFO nova.compute.manager [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Migrating [ 936.455427] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.936s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.455671] env[61728]: DEBUG nova.objects.instance [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lazy-loading 'resources' on Instance uuid c84bfcd0-b145-4675-8b0a-5e8f94f65098 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 936.456974] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.457270] env[61728]: DEBUG nova.compute.manager [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Instance network_info: |[{"id": "4660e4da-30e8-4e51-9aea-a9375732537f", "address": "fa:16:3e:e1:e9:0c", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4660e4da-30", "ovs_interfaceid": "4660e4da-30e8-4e51-9aea-a9375732537f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 936.460174] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:e9:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b80dd748-3d7e-4a23-a38d-9e79a3881452', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4660e4da-30e8-4e51-9aea-a9375732537f', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 936.467551] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Creating folder: Project (f7f61c87df8b485b9a6370032dd7af37). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 936.472803] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a9752e4-3c9f-4f1c-af4e-51bc921e863f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.485997] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Created folder: Project (f7f61c87df8b485b9a6370032dd7af37) in parent group-v121913. [ 936.486234] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Creating folder: Instances. Parent ref: group-v122156. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 936.486473] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d705789-a78e-4dea-9fbc-491a0fc8a52a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.494753] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Created folder: Instances in parent group-v122156. [ 936.494990] env[61728]: DEBUG oslo.service.loopingcall [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.495196] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 936.495395] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-16352096-84bc-4fa4-a234-dbe4f63a05c7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.513692] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 936.513692] env[61728]: value = "task-464513" [ 936.513692] env[61728]: _type = "Task" [ 936.513692] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.520815] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464513, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.702316] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464510, 'name': Rename_Task, 'duration_secs': 0.137108} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.702671] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 936.703024] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-739ce1ef-a112-48b1-b2be-fded42cc55a7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.710049] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 936.710049] env[61728]: value = "task-464514" [ 936.710049] env[61728]: _type = "Task" [ 936.710049] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.717957] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464514, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.738924] env[61728]: DEBUG nova.compute.manager [req-ef31272c-be02-4759-a7ed-fa054461b95b req-5b5e1cd7-426d-47c4-ad2d-9d98cb751bd1 service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Received event network-changed-4660e4da-30e8-4e51-9aea-a9375732537f {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 936.739152] env[61728]: DEBUG nova.compute.manager [req-ef31272c-be02-4759-a7ed-fa054461b95b req-5b5e1cd7-426d-47c4-ad2d-9d98cb751bd1 service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Refreshing instance network info cache due to event network-changed-4660e4da-30e8-4e51-9aea-a9375732537f. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 936.739378] env[61728]: DEBUG oslo_concurrency.lockutils [req-ef31272c-be02-4759-a7ed-fa054461b95b req-5b5e1cd7-426d-47c4-ad2d-9d98cb751bd1 service nova] Acquiring lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.739529] env[61728]: DEBUG oslo_concurrency.lockutils [req-ef31272c-be02-4759-a7ed-fa054461b95b req-5b5e1cd7-426d-47c4-ad2d-9d98cb751bd1 service nova] Acquired lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.739691] env[61728]: DEBUG nova.network.neutron [req-ef31272c-be02-4759-a7ed-fa054461b95b req-5b5e1cd7-426d-47c4-ad2d-9d98cb751bd1 service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Refreshing network info cache for port 4660e4da-30e8-4e51-9aea-a9375732537f {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 936.858056] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "refresh_cache-aebc4c7c-09d6-484f-8ad6-2416b3a41828" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.858056] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "refresh_cache-aebc4c7c-09d6-484f-8ad6-2416b3a41828" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.858183] env[61728]: DEBUG nova.network.neutron [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 936.975364] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.975575] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.975774] env[61728]: DEBUG nova.network.neutron [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 937.023852] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464513, 'name': CreateVM_Task, 'duration_secs': 0.303789} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.026343] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 937.027225] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.027430] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.027705] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 937.027962] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44b85087-d365-4a47-98a6-b04fdaef3967 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.032788] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 937.032788] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52193949-fcb5-9a24-c350-8255f30ef867" [ 937.032788] env[61728]: _type = "Task" [ 937.032788] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.042410] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52193949-fcb5-9a24-c350-8255f30ef867, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.264502] env[61728]: DEBUG oslo_vmware.api [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464514, 'name': PowerOnVM_Task, 'duration_secs': 0.496829} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.264502] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 937.264502] env[61728]: INFO nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Took 10.15 seconds to spawn the instance on the hypervisor. [ 937.264502] env[61728]: DEBUG nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.264502] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b3af40-3100-4484-bde6-2a998af4e577 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.307532] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16c46e0-1e34-49aa-87ce-3b77d0c5e37a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.318370] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41592183-1f5f-4f09-b43c-ecb97741e1ea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.350657] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd191641-b98c-4385-a3bc-c233927d508c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.358954] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8461b499-8342-4f34-b3ea-79a406f96b91 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.373857] env[61728]: DEBUG nova.compute.provider_tree [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.412370] env[61728]: DEBUG nova.network.neutron [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 937.517040] env[61728]: DEBUG nova.network.neutron [req-ef31272c-be02-4759-a7ed-fa054461b95b req-5b5e1cd7-426d-47c4-ad2d-9d98cb751bd1 service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updated VIF entry in instance network info cache for port 4660e4da-30e8-4e51-9aea-a9375732537f. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 937.517385] env[61728]: DEBUG nova.network.neutron [req-ef31272c-be02-4759-a7ed-fa054461b95b req-5b5e1cd7-426d-47c4-ad2d-9d98cb751bd1 service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance_info_cache with network_info: [{"id": "4660e4da-30e8-4e51-9aea-a9375732537f", "address": "fa:16:3e:e1:e9:0c", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4660e4da-30", "ovs_interfaceid": "4660e4da-30e8-4e51-9aea-a9375732537f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.543534] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52193949-fcb5-9a24-c350-8255f30ef867, 'name': SearchDatastore_Task, 'duration_secs': 0.009481} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.543848] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.544099] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 937.544332] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.544479] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.544665] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 937.544929] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b934476-5561-43b1-84f3-81972dccc171 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.557106] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 937.557306] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 937.559695] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bb4fac5-83f9-4833-9efa-04d39c485a6b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.563172] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 937.563172] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522b7d29-a60e-4563-7c34-733a3eca9d45" [ 937.563172] env[61728]: _type = "Task" [ 937.563172] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.571037] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522b7d29-a60e-4563-7c34-733a3eca9d45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.600323] env[61728]: DEBUG nova.network.neutron [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Updating instance_info_cache with network_info: [{"id": "fe024a36-a6a9-44ed-b043-1195882a2ae7", "address": "fa:16:3e:57:76:0a", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe024a36-a6", "ovs_interfaceid": "fe024a36-a6a9-44ed-b043-1195882a2ae7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.694749] env[61728]: DEBUG nova.network.neutron [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating instance_info_cache with network_info: [{"id": "fbf3382b-a343-4cc4-9233-ebf4ac3139a4", "address": "fa:16:3e:e3:49:f4", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbf3382b-a3", "ovs_interfaceid": "fbf3382b-a343-4cc4-9233-ebf4ac3139a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.749757] env[61728]: INFO nova.compute.manager [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Took 46.50 seconds to build instance. [ 937.879067] env[61728]: DEBUG nova.scheduler.client.report [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.020827] env[61728]: DEBUG oslo_concurrency.lockutils [req-ef31272c-be02-4759-a7ed-fa054461b95b req-5b5e1cd7-426d-47c4-ad2d-9d98cb751bd1 service nova] Releasing lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.075221] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522b7d29-a60e-4563-7c34-733a3eca9d45, 'name': SearchDatastore_Task, 'duration_secs': 0.008209} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.076052] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52ad57ad-cca4-4586-abd8-38deeda11d77 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.084089] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 938.084089] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f12a4c-b38f-d1d5-04af-b71fba853a6c" [ 938.084089] env[61728]: _type = "Task" [ 938.084089] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.093301] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f12a4c-b38f-d1d5-04af-b71fba853a6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.103921] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "refresh_cache-aebc4c7c-09d6-484f-8ad6-2416b3a41828" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.104266] env[61728]: DEBUG nova.compute.manager [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Instance network_info: |[{"id": "fe024a36-a6a9-44ed-b043-1195882a2ae7", "address": "fa:16:3e:57:76:0a", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe024a36-a6", "ovs_interfaceid": "fe024a36-a6a9-44ed-b043-1195882a2ae7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 938.104669] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:76:0a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc6d5964-1106-4345-a26d-185dabd4ff0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe024a36-a6a9-44ed-b043-1195882a2ae7', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.112047] env[61728]: DEBUG oslo.service.loopingcall [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.112273] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 938.112502] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c3df6aa8-d29c-4d69-83be-8a8f8d24816a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.133302] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.133302] env[61728]: value = "task-464515" [ 938.133302] env[61728]: _type = "Task" [ 938.133302] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.140932] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464515, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.195945] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.252173] env[61728]: DEBUG oslo_concurrency.lockutils [None req-740db2db-0217-491b-bfb1-9c13c83ed723 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.337s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.266814] env[61728]: DEBUG oslo_concurrency.lockutils [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "0e77ef23-3d1a-4f00-83f9-5b617334f980" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.267115] env[61728]: DEBUG oslo_concurrency.lockutils [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "0e77ef23-3d1a-4f00-83f9-5b617334f980" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.267959] env[61728]: DEBUG oslo_concurrency.lockutils [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "0e77ef23-3d1a-4f00-83f9-5b617334f980-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.267959] env[61728]: DEBUG oslo_concurrency.lockutils [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "0e77ef23-3d1a-4f00-83f9-5b617334f980-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.267959] env[61728]: DEBUG oslo_concurrency.lockutils [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "0e77ef23-3d1a-4f00-83f9-5b617334f980-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.270048] env[61728]: INFO nova.compute.manager [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Terminating instance [ 938.272086] env[61728]: DEBUG nova.compute.manager [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 938.272291] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 938.273131] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036a3773-2c42-4096-a4ef-36d581266535 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.280987] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 938.281267] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f0355466-b216-4ac8-bc7d-84e85fb8c449 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.286965] env[61728]: DEBUG oslo_vmware.api [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 938.286965] env[61728]: value = "task-464516" [ 938.286965] env[61728]: _type = "Task" [ 938.286965] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.295408] env[61728]: DEBUG oslo_vmware.api [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464516, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.318819] env[61728]: DEBUG nova.compute.manager [req-c4a5c0c3-0f1a-466c-aba0-a7086200b040 req-76456667-749f-4cdc-a30e-df29cc9cbe63 service nova] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Received event network-changed-fe024a36-a6a9-44ed-b043-1195882a2ae7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 938.318819] env[61728]: DEBUG nova.compute.manager [req-c4a5c0c3-0f1a-466c-aba0-a7086200b040 req-76456667-749f-4cdc-a30e-df29cc9cbe63 service nova] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Refreshing instance network info cache due to event network-changed-fe024a36-a6a9-44ed-b043-1195882a2ae7. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 938.318886] env[61728]: DEBUG oslo_concurrency.lockutils [req-c4a5c0c3-0f1a-466c-aba0-a7086200b040 req-76456667-749f-4cdc-a30e-df29cc9cbe63 service nova] Acquiring lock "refresh_cache-aebc4c7c-09d6-484f-8ad6-2416b3a41828" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.318999] env[61728]: DEBUG oslo_concurrency.lockutils [req-c4a5c0c3-0f1a-466c-aba0-a7086200b040 req-76456667-749f-4cdc-a30e-df29cc9cbe63 service nova] Acquired lock "refresh_cache-aebc4c7c-09d6-484f-8ad6-2416b3a41828" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.319181] env[61728]: DEBUG nova.network.neutron [req-c4a5c0c3-0f1a-466c-aba0-a7086200b040 req-76456667-749f-4cdc-a30e-df29cc9cbe63 service nova] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Refreshing network info cache for port fe024a36-a6a9-44ed-b043-1195882a2ae7 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 938.384273] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.929s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.388220] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.852s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.388220] env[61728]: DEBUG nova.objects.instance [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lazy-loading 'resources' on Instance uuid 87540771-c864-4d49-8ad3-b6b559841761 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.412599] env[61728]: INFO nova.scheduler.client.report [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Deleted allocations for instance c84bfcd0-b145-4675-8b0a-5e8f94f65098 [ 938.594008] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f12a4c-b38f-d1d5-04af-b71fba853a6c, 'name': SearchDatastore_Task, 'duration_secs': 0.034916} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.594324] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.594603] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 8c5b2380-bd14-451e-9612-2767c2a6adb2/8c5b2380-bd14-451e-9612-2767c2a6adb2.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 938.594865] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5a042bab-69b7-430e-afdd-81165e86429a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.600776] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 938.600776] env[61728]: value = "task-464517" [ 938.600776] env[61728]: _type = "Task" [ 938.600776] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.609843] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464517, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.643898] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464515, 'name': CreateVM_Task, 'duration_secs': 0.401885} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.644132] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 938.644860] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.645067] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.645421] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.645702] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a21b491-205c-4b61-81ce-70fe403ac8aa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.650442] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 938.650442] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528c80c6-1e42-9a1d-5e00-0bbe3f734f5d" [ 938.650442] env[61728]: _type = "Task" [ 938.650442] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.658427] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528c80c6-1e42-9a1d-5e00-0bbe3f734f5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.799497] env[61728]: DEBUG oslo_vmware.api [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464516, 'name': PowerOffVM_Task, 'duration_secs': 0.21515} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.799842] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 938.800039] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 938.800315] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-93e9de83-01de-4708-a3b1-4fb4292f8c50 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.865052] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 938.865507] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 938.865726] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Deleting the datastore file [datastore1] 0e77ef23-3d1a-4f00-83f9-5b617334f980 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.866288] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2b660f2f-db24-4bc7-992f-24d5f5bf2576 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.874978] env[61728]: DEBUG oslo_vmware.api [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 938.874978] env[61728]: value = "task-464519" [ 938.874978] env[61728]: _type = "Task" [ 938.874978] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.886486] env[61728]: DEBUG oslo_vmware.api [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464519, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.921145] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bb21cae-cc66-4e1a-b4b4-0a6870ef152c tempest-VolumesAdminNegativeTest-1125086392 tempest-VolumesAdminNegativeTest-1125086392-project-member] Lock "c84bfcd0-b145-4675-8b0a-5e8f94f65098" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.138s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.085016] env[61728]: DEBUG nova.network.neutron [req-c4a5c0c3-0f1a-466c-aba0-a7086200b040 req-76456667-749f-4cdc-a30e-df29cc9cbe63 service nova] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Updated VIF entry in instance network info cache for port fe024a36-a6a9-44ed-b043-1195882a2ae7. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 939.085450] env[61728]: DEBUG nova.network.neutron [req-c4a5c0c3-0f1a-466c-aba0-a7086200b040 req-76456667-749f-4cdc-a30e-df29cc9cbe63 service nova] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Updating instance_info_cache with network_info: [{"id": "fe024a36-a6a9-44ed-b043-1195882a2ae7", "address": "fa:16:3e:57:76:0a", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe024a36-a6", "ovs_interfaceid": "fe024a36-a6a9-44ed-b043-1195882a2ae7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.116805] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464517, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468385} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.117951] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 8c5b2380-bd14-451e-9612-2767c2a6adb2/8c5b2380-bd14-451e-9612-2767c2a6adb2.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 939.118235] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 939.118504] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-59b9ac30-7e5f-4259-954c-f598376ef63c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.126214] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 939.126214] env[61728]: value = "task-464520" [ 939.126214] env[61728]: _type = "Task" [ 939.126214] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.137527] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464520, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.164235] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528c80c6-1e42-9a1d-5e00-0bbe3f734f5d, 'name': SearchDatastore_Task, 'duration_secs': 0.008519} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.164235] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.164235] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.164512] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.164667] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.164900] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.165528] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c7485ec8-1bba-43be-9e3b-4845931bae6f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.174808] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.174808] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 939.174808] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09e1ab02-f9ed-4a44-b175-2d468080ae9c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.178513] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 939.178513] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d77578-ac35-d6b4-5ead-47ac1768e216" [ 939.178513] env[61728]: _type = "Task" [ 939.178513] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.193156] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d77578-ac35-d6b4-5ead-47ac1768e216, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.284274] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ae2c0c-3e3d-44c3-a448-9b8733d429a2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.292721] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ff9593-789d-4b19-b9a0-bc73f92851d8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.326338] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4cb906-bb9d-4ce9-9392-b9a6e96298fe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.335155] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9c403f-bb40-4054-9410-42fb81bec69a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.349979] env[61728]: DEBUG nova.compute.provider_tree [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.384881] env[61728]: DEBUG oslo_vmware.api [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464519, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.288317} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.385192] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.385386] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 939.385574] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 939.385755] env[61728]: INFO nova.compute.manager [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Took 1.11 seconds to destroy the instance on the hypervisor. [ 939.386016] env[61728]: DEBUG oslo.service.loopingcall [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.386227] env[61728]: DEBUG nova.compute.manager [-] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 939.386343] env[61728]: DEBUG nova.network.neutron [-] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 939.589396] env[61728]: DEBUG oslo_concurrency.lockutils [req-c4a5c0c3-0f1a-466c-aba0-a7086200b040 req-76456667-749f-4cdc-a30e-df29cc9cbe63 service nova] Releasing lock "refresh_cache-aebc4c7c-09d6-484f-8ad6-2416b3a41828" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.639412] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464520, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06339} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.639842] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.640749] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487204cf-4479-4251-b40c-3b4ef2c07b71 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.664530] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 8c5b2380-bd14-451e-9612-2767c2a6adb2/8c5b2380-bd14-451e-9612-2767c2a6adb2.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.665433] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7799fce-e29f-449f-8718-986dc3fd380e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.692385] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d77578-ac35-d6b4-5ead-47ac1768e216, 'name': SearchDatastore_Task, 'duration_secs': 0.009919} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.694458] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 939.694458] env[61728]: value = "task-464521" [ 939.694458] env[61728]: _type = "Task" [ 939.694458] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.694780] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31b948d0-6f66-41d1-81a3-50754a5b1120 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.703621] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 939.703621] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ad1865-5c17-4441-d839-5177f48535ba" [ 939.703621] env[61728]: _type = "Task" [ 939.703621] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.711513] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464521, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.717024] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc9d18a-f3ec-4269-ae11-efb90ce92e68 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.727144] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ad1865-5c17-4441-d839-5177f48535ba, 'name': SearchDatastore_Task, 'duration_secs': 0.010068} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.743584] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.744402] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] aebc4c7c-09d6-484f-8ad6-2416b3a41828/aebc4c7c-09d6-484f-8ad6-2416b3a41828.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 939.744402] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating instance '3415687e-9b8d-42f8-9b18-93951b8ec7da' progress to 0 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 939.748662] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b5ed6e4-db66-49c6-9a07-7e83cf19016d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.755941] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 939.755941] env[61728]: value = "task-464522" [ 939.755941] env[61728]: _type = "Task" [ 939.755941] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.765589] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464522, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.853345] env[61728]: DEBUG nova.scheduler.client.report [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.118562] env[61728]: DEBUG nova.network.neutron [-] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.208807] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464521, 'name': ReconfigVM_Task, 'duration_secs': 0.492735} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.209177] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 8c5b2380-bd14-451e-9612-2767c2a6adb2/8c5b2380-bd14-451e-9612-2767c2a6adb2.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.210413] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fe457100-682b-4e59-b7f9-d2fe6a62d229 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.218019] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 940.218019] env[61728]: value = "task-464523" [ 940.218019] env[61728]: _type = "Task" [ 940.218019] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.225539] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464523, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.255074] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 940.255074] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-558900ed-efa5-4861-99e3-a7d87267c2a9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.260503] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 940.260503] env[61728]: value = "task-464524" [ 940.260503] env[61728]: _type = "Task" [ 940.260503] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.268457] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464522, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.273901] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464524, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.363024] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.973s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.363707] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.007s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.364329] env[61728]: DEBUG nova.objects.instance [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Lazy-loading 'resources' on Instance uuid fef6cb9e-d8c7-45d3-a7f6-e361ff704006 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.406407] env[61728]: INFO nova.scheduler.client.report [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Deleted allocations for instance 87540771-c864-4d49-8ad3-b6b559841761 [ 940.418786] env[61728]: DEBUG nova.compute.manager [req-3f61f740-d0b8-4527-8512-e94e26ffaded req-71d07f66-1104-48a3-98ad-f7f78f8dbffc service nova] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Received event network-vif-deleted-e06b1602-d9c2-459c-bdac-2d1361b2adb9 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 940.624106] env[61728]: INFO nova.compute.manager [-] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Took 1.24 seconds to deallocate network for instance. [ 940.726204] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464523, 'name': Rename_Task, 'duration_secs': 0.252935} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.726545] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 940.726818] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0e37dd7-9f06-45c5-9698-2cd335d3e490 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.733704] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 940.733704] env[61728]: value = "task-464525" [ 940.733704] env[61728]: _type = "Task" [ 940.733704] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.746499] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464525, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.774031] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464522, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.938857} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.775402] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] aebc4c7c-09d6-484f-8ad6-2416b3a41828/aebc4c7c-09d6-484f-8ad6-2416b3a41828.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 940.775784] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.776134] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464524, 'name': PowerOffVM_Task, 'duration_secs': 0.508067} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.776379] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b147d24f-35eb-46c4-b672-8476087965bb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.778344] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 940.778546] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating instance '3415687e-9b8d-42f8-9b18-93951b8ec7da' progress to 17 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 940.788409] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 940.788409] env[61728]: value = "task-464526" [ 940.788409] env[61728]: _type = "Task" [ 940.788409] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.797828] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464526, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.923471] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2ec282fa-83f8-4ee0-880f-5e8c03d79310 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "87540771-c864-4d49-8ad3-b6b559841761" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.224s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.137078] env[61728]: DEBUG oslo_concurrency.lockutils [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.246012] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464525, 'name': PowerOnVM_Task} progress is 79%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.274756] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb24d841-6298-42f3-88cb-c69647cf5e7d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.288333] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.288333] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.288333] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.288333] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.288333] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.288333] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.288333] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.288333] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.288333] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.288333] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.288333] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.294690] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef00dbe0-8ada-4ea2-8d39-28e01b637b4a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.309858] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8278e8-4a6c-4d1c-8530-0a302574910d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.324338] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464526, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073972} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.358846] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.360030] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 941.360030] env[61728]: value = "task-464527" [ 941.360030] env[61728]: _type = "Task" [ 941.360030] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.360453] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2085076-f715-4bb5-868e-15ed079c0c78 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.364181] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ca216ac-79c5-492a-b214-33df4b6d869a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.374729] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.394561] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] aebc4c7c-09d6-484f-8ad6-2416b3a41828/aebc4c7c-09d6-484f-8ad6-2416b3a41828.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.395857] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66415557-65a3-464c-b2a4-3d83c1f12c59 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.400206] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e635685-7653-444d-ad4e-c0bdaa2a66a5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.428521] env[61728]: DEBUG nova.compute.provider_tree [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.431128] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 941.431128] env[61728]: value = "task-464528" [ 941.431128] env[61728]: _type = "Task" [ 941.431128] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.439754] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464528, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.746077] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464525, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.877190] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464527, 'name': ReconfigVM_Task, 'duration_secs': 0.531913} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.877559] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating instance '3415687e-9b8d-42f8-9b18-93951b8ec7da' progress to 33 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 941.933548] env[61728]: DEBUG nova.scheduler.client.report [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.952021] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464528, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.961802] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.961802] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.246758] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464525, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.384823] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 942.385150] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 942.385150] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 942.385915] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 942.385915] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 942.385915] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 942.385915] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 942.386106] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 942.386192] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 942.386323] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 942.386534] env[61728]: DEBUG nova.virt.hardware [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.391763] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Reconfiguring VM instance instance-0000004f to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 942.392082] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0658da19-39bd-4f22-9f2c-49667a992b0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.410539] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 942.410539] env[61728]: value = "task-464529" [ 942.410539] env[61728]: _type = "Task" [ 942.410539] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.419388] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464529, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.446984] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.083s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.450490] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.596s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.450490] env[61728]: INFO nova.compute.claims [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 942.459466] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464528, 'name': ReconfigVM_Task, 'duration_secs': 0.72157} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.461358] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Reconfigured VM instance instance-00000059 to attach disk [datastore1] aebc4c7c-09d6-484f-8ad6-2416b3a41828/aebc4c7c-09d6-484f-8ad6-2416b3a41828.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.464174] env[61728]: INFO nova.compute.manager [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Detaching volume 0835974c-5660-4c88-a603-e2bd917f000b [ 942.466796] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e1cfa57a-9b69-407d-bb8b-5468727a2338 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.474478] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 942.474478] env[61728]: value = "task-464530" [ 942.474478] env[61728]: _type = "Task" [ 942.474478] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.484445] env[61728]: INFO nova.scheduler.client.report [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Deleted allocations for instance fef6cb9e-d8c7-45d3-a7f6-e361ff704006 [ 942.495354] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464530, 'name': Rename_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.533292] env[61728]: INFO nova.virt.block_device [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Attempting to driver detach volume 0835974c-5660-4c88-a603-e2bd917f000b from mountpoint /dev/sdb [ 942.534353] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 942.534353] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122116', 'volume_id': '0835974c-5660-4c88-a603-e2bd917f000b', 'name': 'volume-0835974c-5660-4c88-a603-e2bd917f000b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '06e765e1-85d9-4f6c-8ab5-81735f08cb66', 'attached_at': '', 'detached_at': '', 'volume_id': '0835974c-5660-4c88-a603-e2bd917f000b', 'serial': '0835974c-5660-4c88-a603-e2bd917f000b'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 942.534778] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36cdd1e4-93c1-4054-a050-ddb15c0eebe0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.559365] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a266c15-ed38-46cd-be7b-f20ecdc068d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.568044] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc243e5-0746-4a42-8edb-7c4a9b21c3f0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.595009] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588c93d0-c2de-4ac0-821d-ff84792cc5ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.612861] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] The volume has not been displaced from its original location: [datastore1] volume-0835974c-5660-4c88-a603-e2bd917f000b/volume-0835974c-5660-4c88-a603-e2bd917f000b.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 942.618697] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Reconfiguring VM instance instance-0000003b to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 942.618697] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d093676-8e62-40fc-9f26-84939aeda736 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.639749] env[61728]: DEBUG oslo_vmware.api [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 942.639749] env[61728]: value = "task-464531" [ 942.639749] env[61728]: _type = "Task" [ 942.639749] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.648780] env[61728]: DEBUG oslo_vmware.api [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464531, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.746564] env[61728]: DEBUG oslo_vmware.api [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464525, 'name': PowerOnVM_Task, 'duration_secs': 1.603086} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.746858] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 942.747080] env[61728]: INFO nova.compute.manager [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Took 9.25 seconds to spawn the instance on the hypervisor. [ 942.747288] env[61728]: DEBUG nova.compute.manager [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 942.748049] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e84fa01-32d9-4db8-80cb-a9936ef07202 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.924053] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464529, 'name': ReconfigVM_Task, 'duration_secs': 0.162681} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.924053] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Reconfigured VM instance instance-0000004f to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 942.924053] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-442a4b7a-05e3-4ec9-8ae0-fbb77f58d528 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.951471] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 3415687e-9b8d-42f8-9b18-93951b8ec7da/3415687e-9b8d-42f8-9b18-93951b8ec7da.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.951967] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa8ff585-a37d-47f2-9106-2baeebccbc54 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.978351] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 942.978351] env[61728]: value = "task-464532" [ 942.978351] env[61728]: _type = "Task" [ 942.978351] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.991971] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464532, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.995659] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464530, 'name': Rename_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.005094] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c941c0b0-20a4-4b50-8505-c5c4eb2c6d2b tempest-ServerTagsTestJSON-1845761596 tempest-ServerTagsTestJSON-1845761596-project-member] Lock "fef6cb9e-d8c7-45d3-a7f6-e361ff704006" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.664s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.150088] env[61728]: DEBUG oslo_vmware.api [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464531, 'name': ReconfigVM_Task, 'duration_secs': 0.252371} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.150380] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Reconfigured VM instance instance-0000003b to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 943.155638] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4da30251-da35-4e38-ab4c-660100c022ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.175252] env[61728]: DEBUG oslo_vmware.api [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 943.175252] env[61728]: value = "task-464533" [ 943.175252] env[61728]: _type = "Task" [ 943.175252] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.184866] env[61728]: DEBUG oslo_vmware.api [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464533, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.264376] env[61728]: INFO nova.compute.manager [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Took 44.85 seconds to build instance. [ 943.496709] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464532, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.497381] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464530, 'name': Rename_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.686774] env[61728]: DEBUG oslo_vmware.api [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464533, 'name': ReconfigVM_Task, 'duration_secs': 0.140045} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.690382] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122116', 'volume_id': '0835974c-5660-4c88-a603-e2bd917f000b', 'name': 'volume-0835974c-5660-4c88-a603-e2bd917f000b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '06e765e1-85d9-4f6c-8ab5-81735f08cb66', 'attached_at': '', 'detached_at': '', 'volume_id': '0835974c-5660-4c88-a603-e2bd917f000b', 'serial': '0835974c-5660-4c88-a603-e2bd917f000b'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 943.768145] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c3511996-5268-4444-b2f7-6d82248f13fe tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.456s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.836465] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76aa7c60-5a89-4c95-ac02-a85cf9c99495 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.846170] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68dda26d-c756-44ed-94ca-25fea8801e1c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.893703] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e0fcd89-92ea-4923-8dea-ce84d21cdc9b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.906233] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70dcdd4-c5bb-4689-89f0-1222e3284bb2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.928821] env[61728]: DEBUG nova.compute.provider_tree [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.995815] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464530, 'name': Rename_Task, 'duration_secs': 1.132465} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.996093] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464532, 'name': ReconfigVM_Task, 'duration_secs': 0.872652} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.996345] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 943.996608] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 3415687e-9b8d-42f8-9b18-93951b8ec7da/3415687e-9b8d-42f8-9b18-93951b8ec7da.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 943.996883] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating instance '3415687e-9b8d-42f8-9b18-93951b8ec7da' progress to 50 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 944.000144] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55f6250b-0f14-486f-a990-d17667e11fb7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.007478] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 944.007478] env[61728]: value = "task-464534" [ 944.007478] env[61728]: _type = "Task" [ 944.007478] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.015219] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464534, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.250930] env[61728]: DEBUG nova.objects.instance [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lazy-loading 'flavor' on Instance uuid 06e765e1-85d9-4f6c-8ab5-81735f08cb66 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.433311] env[61728]: DEBUG nova.scheduler.client.report [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.506146] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d69edb4c-4648-4455-b845-bc6209bc370b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.542820] env[61728]: DEBUG oslo_vmware.api [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464534, 'name': PowerOnVM_Task, 'duration_secs': 0.459023} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.544252] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c107fd07-9173-4916-b942-f2face4cf6f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.546769] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 944.548074] env[61728]: INFO nova.compute.manager [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Took 8.56 seconds to spawn the instance on the hypervisor. [ 944.548074] env[61728]: DEBUG nova.compute.manager [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.548175] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa692c5-77ca-4004-9cb3-b4d9d2298749 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.568982] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating instance '3415687e-9b8d-42f8-9b18-93951b8ec7da' progress to 67 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 944.752651] env[61728]: DEBUG nova.compute.manager [req-98b581bd-0fe7-4d1c-9fb2-862fb4a18935 req-a308e88b-5c65-4a25-a391-59f8a20a587f service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Received event network-changed-4660e4da-30e8-4e51-9aea-a9375732537f {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 944.753256] env[61728]: DEBUG nova.compute.manager [req-98b581bd-0fe7-4d1c-9fb2-862fb4a18935 req-a308e88b-5c65-4a25-a391-59f8a20a587f service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Refreshing instance network info cache due to event network-changed-4660e4da-30e8-4e51-9aea-a9375732537f. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 944.753597] env[61728]: DEBUG oslo_concurrency.lockutils [req-98b581bd-0fe7-4d1c-9fb2-862fb4a18935 req-a308e88b-5c65-4a25-a391-59f8a20a587f service nova] Acquiring lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.753936] env[61728]: DEBUG oslo_concurrency.lockutils [req-98b581bd-0fe7-4d1c-9fb2-862fb4a18935 req-a308e88b-5c65-4a25-a391-59f8a20a587f service nova] Acquired lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.754188] env[61728]: DEBUG nova.network.neutron [req-98b581bd-0fe7-4d1c-9fb2-862fb4a18935 req-a308e88b-5c65-4a25-a391-59f8a20a587f service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Refreshing network info cache for port 4660e4da-30e8-4e51-9aea-a9375732537f {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 944.945821] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.945821] env[61728]: DEBUG nova.compute.manager [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 944.946547] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 33.195s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.946748] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.946911] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61728) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 944.947320] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.779s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.948767] env[61728]: INFO nova.compute.claims [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.952652] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371d8cce-8aca-4575-a014-1fae721077df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.964686] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec248c7-6b73-49b9-91d3-cffc3a2871a0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.988330] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974bb431-1f80-445a-b10d-9581bc25fdab {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.996948] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190bacd0-2a0a-4e9a-bac1-e2bf8ed515d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.033045] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180529MB free_disk=114GB free_vcpus=48 pci_devices=None {{(pid=61728) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 945.033484] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.093694] env[61728]: INFO nova.compute.manager [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Took 46.43 seconds to build instance. [ 945.122581] env[61728]: DEBUG nova.network.neutron [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Port fbf3382b-a343-4cc4-9233-ebf4ac3139a4 binding to destination host cpu-1 is already ACTIVE {{(pid=61728) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 945.262338] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ad32ad07-cf20-4202-a09d-da3786b21355 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.301s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.453579] env[61728]: DEBUG nova.compute.utils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 945.459111] env[61728]: DEBUG nova.compute.manager [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 945.459352] env[61728]: DEBUG nova.network.neutron [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 945.524172] env[61728]: DEBUG nova.network.neutron [req-98b581bd-0fe7-4d1c-9fb2-862fb4a18935 req-a308e88b-5c65-4a25-a391-59f8a20a587f service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updated VIF entry in instance network info cache for port 4660e4da-30e8-4e51-9aea-a9375732537f. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 945.524172] env[61728]: DEBUG nova.network.neutron [req-98b581bd-0fe7-4d1c-9fb2-862fb4a18935 req-a308e88b-5c65-4a25-a391-59f8a20a587f service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance_info_cache with network_info: [{"id": "4660e4da-30e8-4e51-9aea-a9375732537f", "address": "fa:16:3e:e1:e9:0c", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4660e4da-30", "ovs_interfaceid": "4660e4da-30e8-4e51-9aea-a9375732537f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.567330] env[61728]: DEBUG nova.policy [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e51efcf23db4d5496edbc3b7104b4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32bb35e1dfce40e48be08bb568d3f2b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 945.596103] env[61728]: DEBUG oslo_concurrency.lockutils [None req-40e91c02-ca31-463b-a85a-3de98dca8cf4 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "aebc4c7c-09d6-484f-8ad6-2416b3a41828" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.484s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.962444] env[61728]: DEBUG nova.compute.manager [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 946.028127] env[61728]: DEBUG oslo_concurrency.lockutils [req-98b581bd-0fe7-4d1c-9fb2-862fb4a18935 req-a308e88b-5c65-4a25-a391-59f8a20a587f service nova] Releasing lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.097964] env[61728]: DEBUG nova.network.neutron [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Successfully created port: c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 946.154386] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "3415687e-9b8d-42f8-9b18-93951b8ec7da-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.154386] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "3415687e-9b8d-42f8-9b18-93951b8ec7da-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.154386] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "3415687e-9b8d-42f8-9b18-93951b8ec7da-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.299966] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "aebc4c7c-09d6-484f-8ad6-2416b3a41828" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.300429] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "aebc4c7c-09d6-484f-8ad6-2416b3a41828" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.300512] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "aebc4c7c-09d6-484f-8ad6-2416b3a41828-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.300671] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "aebc4c7c-09d6-484f-8ad6-2416b3a41828-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.300848] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "aebc4c7c-09d6-484f-8ad6-2416b3a41828-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.302957] env[61728]: INFO nova.compute.manager [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Terminating instance [ 946.304781] env[61728]: DEBUG nova.compute.manager [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 946.304988] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 946.305945] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a53b9e7-79ed-4539-b9ab-def8a849619b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.316388] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 946.316649] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-18a5f38b-bdeb-4b12-aa2f-eada2f0809f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.323212] env[61728]: DEBUG oslo_vmware.api [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 946.323212] env[61728]: value = "task-464535" [ 946.323212] env[61728]: _type = "Task" [ 946.323212] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.332416] env[61728]: DEBUG oslo_vmware.api [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464535, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.404265] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f5fcb6-cb92-40bd-900d-7bf2e6eb850f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.411582] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a72fa754-ec15-4043-9cdf-c176a34c5e11 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.443736] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d21ae962-4440-460f-a3e1-b4aeb0ebf483 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.451398] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f5b6020-f130-4033-aefc-07583b494b2f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.466372] env[61728]: DEBUG nova.compute.provider_tree [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.476021] env[61728]: DEBUG oslo_concurrency.lockutils [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.476285] env[61728]: DEBUG oslo_concurrency.lockutils [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.476483] env[61728]: DEBUG oslo_concurrency.lockutils [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.476669] env[61728]: DEBUG oslo_concurrency.lockutils [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.476840] env[61728]: DEBUG oslo_concurrency.lockutils [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.478885] env[61728]: INFO nova.compute.manager [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Terminating instance [ 946.480569] env[61728]: DEBUG nova.compute.manager [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 946.480795] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 946.481830] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c3b112-623d-4e17-9172-1226cc057f3f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.488763] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 946.488871] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5a9cc8a-05aa-4f90-a715-e82c8ba86c8a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.494423] env[61728]: DEBUG oslo_vmware.api [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 946.494423] env[61728]: value = "task-464536" [ 946.494423] env[61728]: _type = "Task" [ 946.494423] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.503132] env[61728]: DEBUG oslo_vmware.api [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464536, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.833019] env[61728]: DEBUG oslo_vmware.api [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464535, 'name': PowerOffVM_Task, 'duration_secs': 0.175596} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.833357] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 946.833546] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 946.833810] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e28e0bbd-2e35-4951-87b0-c29d61c2a0bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.892541] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 946.892780] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 946.892969] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleting the datastore file [datastore1] aebc4c7c-09d6-484f-8ad6-2416b3a41828 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.893266] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0dc7ef1e-8435-4ba4-bb7e-e23bcc2e1e7b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.898986] env[61728]: DEBUG oslo_vmware.api [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 946.898986] env[61728]: value = "task-464538" [ 946.898986] env[61728]: _type = "Task" [ 946.898986] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.907214] env[61728]: DEBUG oslo_vmware.api [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464538, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.969234] env[61728]: DEBUG nova.scheduler.client.report [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.979850] env[61728]: DEBUG nova.compute.manager [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 947.007501] env[61728]: DEBUG oslo_vmware.api [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464536, 'name': PowerOffVM_Task, 'duration_secs': 0.197215} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.009648] env[61728]: DEBUG nova.virt.hardware [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 947.009893] env[61728]: DEBUG nova.virt.hardware [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 947.010073] env[61728]: DEBUG nova.virt.hardware [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 947.010267] env[61728]: DEBUG nova.virt.hardware [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 947.010423] env[61728]: DEBUG nova.virt.hardware [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 947.010596] env[61728]: DEBUG nova.virt.hardware [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 947.010822] env[61728]: DEBUG nova.virt.hardware [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 947.011014] env[61728]: DEBUG nova.virt.hardware [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 947.011215] env[61728]: DEBUG nova.virt.hardware [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 947.011375] env[61728]: DEBUG nova.virt.hardware [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 947.011952] env[61728]: DEBUG nova.virt.hardware [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 947.012073] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 947.012214] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 947.012999] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18d49b2-12c4-4acf-889c-6e5db0208382 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.015490] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-552a07db-d992-4946-89bc-61a63ebd9645 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.023060] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2318e3-b89f-43df-933f-1acaef9cf2d2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.078359] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 947.078554] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 947.079013] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Deleting the datastore file [datastore1] 06e765e1-85d9-4f6c-8ab5-81735f08cb66 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 947.079013] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5810942a-a07f-42b7-99e6-36d9488abb5f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.086478] env[61728]: DEBUG oslo_vmware.api [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 947.086478] env[61728]: value = "task-464540" [ 947.086478] env[61728]: _type = "Task" [ 947.086478] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.095137] env[61728]: DEBUG oslo_vmware.api [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464540, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.186477] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.186706] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.186912] env[61728]: DEBUG nova.network.neutron [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 947.408947] env[61728]: DEBUG oslo_vmware.api [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464538, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150698} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.409342] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 947.409540] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 947.409724] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 947.409906] env[61728]: INFO nova.compute.manager [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Took 1.10 seconds to destroy the instance on the hypervisor. [ 947.410178] env[61728]: DEBUG oslo.service.loopingcall [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.410387] env[61728]: DEBUG nova.compute.manager [-] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 947.410483] env[61728]: DEBUG nova.network.neutron [-] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 947.476868] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.529s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.478151] env[61728]: DEBUG nova.compute.manager [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 947.483805] env[61728]: DEBUG oslo_concurrency.lockutils [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.403s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.484382] env[61728]: DEBUG nova.objects.instance [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lazy-loading 'resources' on Instance uuid 06109957-2c3b-404e-a43e-dd34ece39096 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.536942] env[61728]: DEBUG nova.compute.manager [req-eda2dbc6-4a72-48ae-bd70-9c2a0521612a req-bf66abcf-37ad-4596-9bd6-beea22680a7f service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Received event network-vif-plugged-c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 947.537300] env[61728]: DEBUG oslo_concurrency.lockutils [req-eda2dbc6-4a72-48ae-bd70-9c2a0521612a req-bf66abcf-37ad-4596-9bd6-beea22680a7f service nova] Acquiring lock "a70d1c37-e792-4168-b7e1-9418b8cb7818-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.537544] env[61728]: DEBUG oslo_concurrency.lockutils [req-eda2dbc6-4a72-48ae-bd70-9c2a0521612a req-bf66abcf-37ad-4596-9bd6-beea22680a7f service nova] Lock "a70d1c37-e792-4168-b7e1-9418b8cb7818-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.537755] env[61728]: DEBUG oslo_concurrency.lockutils [req-eda2dbc6-4a72-48ae-bd70-9c2a0521612a req-bf66abcf-37ad-4596-9bd6-beea22680a7f service nova] Lock "a70d1c37-e792-4168-b7e1-9418b8cb7818-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.537963] env[61728]: DEBUG nova.compute.manager [req-eda2dbc6-4a72-48ae-bd70-9c2a0521612a req-bf66abcf-37ad-4596-9bd6-beea22680a7f service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] No waiting events found dispatching network-vif-plugged-c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 947.538191] env[61728]: WARNING nova.compute.manager [req-eda2dbc6-4a72-48ae-bd70-9c2a0521612a req-bf66abcf-37ad-4596-9bd6-beea22680a7f service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Received unexpected event network-vif-plugged-c8407eac-4337-4ae3-a071-08f48bf87248 for instance with vm_state building and task_state spawning. [ 947.601889] env[61728]: DEBUG oslo_vmware.api [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464540, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128579} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.605016] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 947.605016] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 947.605016] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 947.605016] env[61728]: INFO nova.compute.manager [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Took 1.12 seconds to destroy the instance on the hypervisor. [ 947.605016] env[61728]: DEBUG oslo.service.loopingcall [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.605016] env[61728]: DEBUG nova.compute.manager [-] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 947.605016] env[61728]: DEBUG nova.network.neutron [-] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 947.711797] env[61728]: DEBUG nova.network.neutron [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Successfully updated port: c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 947.738657] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "f9f07464-9f43-43fd-8895-2673861747fa" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.740345] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "f9f07464-9f43-43fd-8895-2673861747fa" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.820493] env[61728]: DEBUG nova.compute.manager [req-f5eba6bb-72c5-4abf-9d92-154f0637171c req-07270cce-5f3b-4e1e-805f-e8dab081a34b service nova] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Received event network-vif-deleted-fe024a36-a6a9-44ed-b043-1195882a2ae7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 947.822232] env[61728]: INFO nova.compute.manager [req-f5eba6bb-72c5-4abf-9d92-154f0637171c req-07270cce-5f3b-4e1e-805f-e8dab081a34b service nova] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Neutron deleted interface fe024a36-a6a9-44ed-b043-1195882a2ae7; detaching it from the instance and deleting it from the info cache [ 947.822232] env[61728]: DEBUG nova.network.neutron [req-f5eba6bb-72c5-4abf-9d92-154f0637171c req-07270cce-5f3b-4e1e-805f-e8dab081a34b service nova] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.986017] env[61728]: DEBUG nova.compute.utils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 947.987185] env[61728]: DEBUG nova.compute.manager [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 947.991027] env[61728]: DEBUG nova.network.neutron [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 948.061123] env[61728]: DEBUG nova.policy [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '091abd20ebfe450b9283f5c7df1c7993', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c78292171414530a1caba3dfd0f620f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 948.063471] env[61728]: DEBUG nova.network.neutron [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating instance_info_cache with network_info: [{"id": "fbf3382b-a343-4cc4-9233-ebf4ac3139a4", "address": "fa:16:3e:e3:49:f4", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbf3382b-a3", "ovs_interfaceid": "fbf3382b-a343-4cc4-9233-ebf4ac3139a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.216068] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.216068] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.216174] env[61728]: DEBUG nova.network.neutron [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 948.220747] env[61728]: DEBUG nova.network.neutron [-] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.244748] env[61728]: DEBUG nova.compute.utils [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 948.324802] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e4d9f86b-5eb5-4852-812a-de70542b8938 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.335676] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb97984-6537-4aba-8f7e-1dd381a322e5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.369166] env[61728]: DEBUG nova.compute.manager [req-f5eba6bb-72c5-4abf-9d92-154f0637171c req-07270cce-5f3b-4e1e-805f-e8dab081a34b service nova] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Detach interface failed, port_id=fe024a36-a6a9-44ed-b043-1195882a2ae7, reason: Instance aebc4c7c-09d6-484f-8ad6-2416b3a41828 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 948.383758] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d827305a-4ceb-4d96-a432-7392403c9078 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.394283] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c4a442-4020-4035-bc9e-5c0f7f0ddd32 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.426753] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3858fa22-f1f5-4c94-8843-867a6bf586f7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.434399] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac4e5e4e-b1a9-4245-8e89-f9922a41310c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.447895] env[61728]: DEBUG nova.compute.provider_tree [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.449920] env[61728]: DEBUG nova.network.neutron [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Successfully created port: 814bcd52-6c0c-4efd-ae00-1cfb3375cd63 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.492702] env[61728]: DEBUG nova.compute.manager [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 948.567234] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.724124] env[61728]: INFO nova.compute.manager [-] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Took 1.31 seconds to deallocate network for instance. [ 948.745144] env[61728]: DEBUG nova.network.neutron [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 948.750217] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "f9f07464-9f43-43fd-8895-2673861747fa" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.869938] env[61728]: DEBUG nova.network.neutron [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updating instance_info_cache with network_info: [{"id": "c8407eac-4337-4ae3-a071-08f48bf87248", "address": "fa:16:3e:2d:a6:d6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8407eac-43", "ovs_interfaceid": "c8407eac-4337-4ae3-a071-08f48bf87248", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.877329] env[61728]: DEBUG nova.network.neutron [-] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.953178] env[61728]: DEBUG nova.scheduler.client.report [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.092376] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5908a8-3e43-49de-b258-a7a81c0d5a07 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.111603] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c6d99b-a804-4956-be17-6a7ae4ea67e4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.118328] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating instance '3415687e-9b8d-42f8-9b18-93951b8ec7da' progress to 83 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 949.231215] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.372910] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.373273] env[61728]: DEBUG nova.compute.manager [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Instance network_info: |[{"id": "c8407eac-4337-4ae3-a071-08f48bf87248", "address": "fa:16:3e:2d:a6:d6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8407eac-43", "ovs_interfaceid": "c8407eac-4337-4ae3-a071-08f48bf87248", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 949.373707] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2d:a6:d6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c8407eac-4337-4ae3-a071-08f48bf87248', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 949.381302] env[61728]: DEBUG oslo.service.loopingcall [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 949.381526] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 949.382868] env[61728]: INFO nova.compute.manager [-] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Took 1.78 seconds to deallocate network for instance. [ 949.382868] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1a28bc1b-a54a-4080-816e-b257b43fbe71 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.408977] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 949.408977] env[61728]: value = "task-464541" [ 949.408977] env[61728]: _type = "Task" [ 949.408977] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.422023] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464541, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.459425] env[61728]: DEBUG oslo_concurrency.lockutils [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.975s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.462442] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.352s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.462969] env[61728]: DEBUG nova.objects.instance [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lazy-loading 'pci_requests' on Instance uuid 3bfefcfc-db97-4a9d-86cb-9fb1d8158863 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.481451] env[61728]: INFO nova.scheduler.client.report [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleted allocations for instance 06109957-2c3b-404e-a43e-dd34ece39096 [ 949.502381] env[61728]: DEBUG nova.compute.manager [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 949.527366] env[61728]: DEBUG nova.virt.hardware [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.527667] env[61728]: DEBUG nova.virt.hardware [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.527871] env[61728]: DEBUG nova.virt.hardware [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.528211] env[61728]: DEBUG nova.virt.hardware [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.528415] env[61728]: DEBUG nova.virt.hardware [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.528596] env[61728]: DEBUG nova.virt.hardware [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.528832] env[61728]: DEBUG nova.virt.hardware [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.529031] env[61728]: DEBUG nova.virt.hardware [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.529241] env[61728]: DEBUG nova.virt.hardware [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.529434] env[61728]: DEBUG nova.virt.hardware [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.529692] env[61728]: DEBUG nova.virt.hardware [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.530567] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdfcf0aa-70b2-40a3-acfc-42cdb8eae880 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.539987] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221079c2-10dd-4591-9118-6928750cad62 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.566674] env[61728]: DEBUG nova.compute.manager [req-d5aecf9f-1510-4f2b-9539-eec07b75496d req-cd48a51d-b533-41b1-809d-b701f3630a82 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Received event network-changed-c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 949.566914] env[61728]: DEBUG nova.compute.manager [req-d5aecf9f-1510-4f2b-9539-eec07b75496d req-cd48a51d-b533-41b1-809d-b701f3630a82 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Refreshing instance network info cache due to event network-changed-c8407eac-4337-4ae3-a071-08f48bf87248. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 949.567173] env[61728]: DEBUG oslo_concurrency.lockutils [req-d5aecf9f-1510-4f2b-9539-eec07b75496d req-cd48a51d-b533-41b1-809d-b701f3630a82 service nova] Acquiring lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.567344] env[61728]: DEBUG oslo_concurrency.lockutils [req-d5aecf9f-1510-4f2b-9539-eec07b75496d req-cd48a51d-b533-41b1-809d-b701f3630a82 service nova] Acquired lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.567521] env[61728]: DEBUG nova.network.neutron [req-d5aecf9f-1510-4f2b-9539-eec07b75496d req-cd48a51d-b533-41b1-809d-b701f3630a82 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Refreshing network info cache for port c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 949.625381] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 949.625679] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-64b5672c-a3f8-42a7-afd2-04a7f70ac1db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.632368] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 949.632368] env[61728]: value = "task-464542" [ 949.632368] env[61728]: _type = "Task" [ 949.632368] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.640488] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464542, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.809285] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "f9f07464-9f43-43fd-8895-2673861747fa" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.809647] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "f9f07464-9f43-43fd-8895-2673861747fa" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.810234] env[61728]: INFO nova.compute.manager [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Attaching volume fe562e41-06ae-47c9-9eda-4f075c89cd34 to /dev/sdb [ 949.850044] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d77d2d-cba9-41ee-991f-a419ab3d8db1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.857721] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d99ad2b2-e999-488e-94f7-28511dd00688 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.862512] env[61728]: DEBUG nova.compute.manager [req-8d2c93a3-6a2c-45ae-b3f5-ee79d9a05aca req-438a8f69-153e-445b-9bee-8597d0279725 service nova] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Received event network-vif-deleted-33a9ae57-3181-42d6-9a7f-a66bf8103138 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 949.862736] env[61728]: DEBUG nova.compute.manager [req-8d2c93a3-6a2c-45ae-b3f5-ee79d9a05aca req-438a8f69-153e-445b-9bee-8597d0279725 service nova] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Received event network-vif-plugged-814bcd52-6c0c-4efd-ae00-1cfb3375cd63 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 949.862947] env[61728]: DEBUG oslo_concurrency.lockutils [req-8d2c93a3-6a2c-45ae-b3f5-ee79d9a05aca req-438a8f69-153e-445b-9bee-8597d0279725 service nova] Acquiring lock "0af88a53-bbe7-443e-a7dc-734e483bb730-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.863412] env[61728]: DEBUG oslo_concurrency.lockutils [req-8d2c93a3-6a2c-45ae-b3f5-ee79d9a05aca req-438a8f69-153e-445b-9bee-8597d0279725 service nova] Lock "0af88a53-bbe7-443e-a7dc-734e483bb730-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.863597] env[61728]: DEBUG oslo_concurrency.lockutils [req-8d2c93a3-6a2c-45ae-b3f5-ee79d9a05aca req-438a8f69-153e-445b-9bee-8597d0279725 service nova] Lock "0af88a53-bbe7-443e-a7dc-734e483bb730-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.863977] env[61728]: DEBUG nova.compute.manager [req-8d2c93a3-6a2c-45ae-b3f5-ee79d9a05aca req-438a8f69-153e-445b-9bee-8597d0279725 service nova] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] No waiting events found dispatching network-vif-plugged-814bcd52-6c0c-4efd-ae00-1cfb3375cd63 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 949.864178] env[61728]: WARNING nova.compute.manager [req-8d2c93a3-6a2c-45ae-b3f5-ee79d9a05aca req-438a8f69-153e-445b-9bee-8597d0279725 service nova] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Received unexpected event network-vif-plugged-814bcd52-6c0c-4efd-ae00-1cfb3375cd63 for instance with vm_state building and task_state spawning. [ 949.875191] env[61728]: DEBUG nova.virt.block_device [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating existing volume attachment record: 978d7031-d136-46b9-aa5f-5b45b95accd6 {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 949.905331] env[61728]: DEBUG oslo_concurrency.lockutils [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.922041] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464541, 'name': CreateVM_Task, 'duration_secs': 0.347562} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.922718] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 949.923764] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.924016] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.924459] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 949.924844] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d06c68ee-8067-4d48-ad48-45f61bfaa074 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.931653] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 949.931653] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5233993b-d851-216e-be63-fbc44951857b" [ 949.931653] env[61728]: _type = "Task" [ 949.931653] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.932674] env[61728]: DEBUG nova.network.neutron [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Successfully updated port: 814bcd52-6c0c-4efd-ae00-1cfb3375cd63 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 949.942422] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5233993b-d851-216e-be63-fbc44951857b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.967961] env[61728]: DEBUG nova.objects.instance [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lazy-loading 'numa_topology' on Instance uuid 3bfefcfc-db97-4a9d-86cb-9fb1d8158863 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.989383] env[61728]: DEBUG oslo_concurrency.lockutils [None req-84c71a80-2f5e-4f78-9fc9-c4c02549d95f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "06109957-2c3b-404e-a43e-dd34ece39096" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.304s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.147166] env[61728]: DEBUG oslo_vmware.api [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464542, 'name': PowerOnVM_Task, 'duration_secs': 0.398529} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.152344] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 950.152714] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e7948d93-4896-498a-a2c7-a7bff94dd30a tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating instance '3415687e-9b8d-42f8-9b18-93951b8ec7da' progress to 100 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 950.303309] env[61728]: DEBUG nova.network.neutron [req-d5aecf9f-1510-4f2b-9539-eec07b75496d req-cd48a51d-b533-41b1-809d-b701f3630a82 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updated VIF entry in instance network info cache for port c8407eac-4337-4ae3-a071-08f48bf87248. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 950.303716] env[61728]: DEBUG nova.network.neutron [req-d5aecf9f-1510-4f2b-9539-eec07b75496d req-cd48a51d-b533-41b1-809d-b701f3630a82 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updating instance_info_cache with network_info: [{"id": "c8407eac-4337-4ae3-a071-08f48bf87248", "address": "fa:16:3e:2d:a6:d6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8407eac-43", "ovs_interfaceid": "c8407eac-4337-4ae3-a071-08f48bf87248", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.438431] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "refresh_cache-0af88a53-bbe7-443e-a7dc-734e483bb730" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.438563] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "refresh_cache-0af88a53-bbe7-443e-a7dc-734e483bb730" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.438718] env[61728]: DEBUG nova.network.neutron [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 950.443741] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5233993b-d851-216e-be63-fbc44951857b, 'name': SearchDatastore_Task, 'duration_secs': 0.011132} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.444300] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.444848] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 950.444848] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.445008] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.445125] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 950.445350] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-513a6b21-1c9b-4f40-acc9-67fd310430d2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.455621] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.455827] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 950.456858] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2111f3a0-d4be-4b0a-933d-fa55b9b6fedc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.462259] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 950.462259] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529b3009-0cac-71b3-c2b8-dd4670edc55e" [ 950.462259] env[61728]: _type = "Task" [ 950.462259] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.469801] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529b3009-0cac-71b3-c2b8-dd4670edc55e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.470229] env[61728]: INFO nova.compute.claims [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.807046] env[61728]: DEBUG oslo_concurrency.lockutils [req-d5aecf9f-1510-4f2b-9539-eec07b75496d req-cd48a51d-b533-41b1-809d-b701f3630a82 service nova] Releasing lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.973238] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529b3009-0cac-71b3-c2b8-dd4670edc55e, 'name': SearchDatastore_Task, 'duration_secs': 0.011036} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.976149] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ea4f4d7-fba6-43d2-89a7-778979b17eaa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.979289] env[61728]: DEBUG nova.network.neutron [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 950.984422] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 950.984422] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5261c43b-df33-03a3-490c-683533ab37cc" [ 950.984422] env[61728]: _type = "Task" [ 950.984422] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.993584] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5261c43b-df33-03a3-490c-683533ab37cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.126313] env[61728]: DEBUG nova.network.neutron [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Updating instance_info_cache with network_info: [{"id": "814bcd52-6c0c-4efd-ae00-1cfb3375cd63", "address": "fa:16:3e:6f:a5:67", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap814bcd52-6c", "ovs_interfaceid": "814bcd52-6c0c-4efd-ae00-1cfb3375cd63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.495224] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5261c43b-df33-03a3-490c-683533ab37cc, 'name': SearchDatastore_Task, 'duration_secs': 0.013402} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.495597] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.495816] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a70d1c37-e792-4168-b7e1-9418b8cb7818/a70d1c37-e792-4168-b7e1-9418b8cb7818.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 951.496124] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1f37acb5-5db1-4ba4-983e-f474971273cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.506017] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 951.506017] env[61728]: value = "task-464546" [ 951.506017] env[61728]: _type = "Task" [ 951.506017] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.514611] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464546, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.630400] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "refresh_cache-0af88a53-bbe7-443e-a7dc-734e483bb730" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.630400] env[61728]: DEBUG nova.compute.manager [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Instance network_info: |[{"id": "814bcd52-6c0c-4efd-ae00-1cfb3375cd63", "address": "fa:16:3e:6f:a5:67", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap814bcd52-6c", "ovs_interfaceid": "814bcd52-6c0c-4efd-ae00-1cfb3375cd63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 951.630400] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:a5:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '814bcd52-6c0c-4efd-ae00-1cfb3375cd63', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.638085] env[61728]: DEBUG oslo.service.loopingcall [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.641083] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 951.642035] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b8b7f59-623e-4d98-91ee-743860c4fae6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.664695] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.664695] env[61728]: value = "task-464547" [ 951.664695] env[61728]: _type = "Task" [ 951.664695] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.680580] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464547, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.817165] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64067b9-6185-4eaf-98e7-c1cc558c9e08 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.825044] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c204fe-9de1-4054-ab60-312a5c31cc97 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.855433] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6001d3-b541-4ad6-aedb-1bf9c5aaf4e9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.863026] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84fffa6c-2667-4859-a13c-360f855de8ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.877530] env[61728]: DEBUG nova.compute.provider_tree [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.890972] env[61728]: DEBUG nova.compute.manager [req-f7857d60-43da-4db6-8f59-9756d4229107 req-3ec051fb-ee67-45a6-9ed0-cacae495e546 service nova] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Received event network-changed-814bcd52-6c0c-4efd-ae00-1cfb3375cd63 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 951.890972] env[61728]: DEBUG nova.compute.manager [req-f7857d60-43da-4db6-8f59-9756d4229107 req-3ec051fb-ee67-45a6-9ed0-cacae495e546 service nova] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Refreshing instance network info cache due to event network-changed-814bcd52-6c0c-4efd-ae00-1cfb3375cd63. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 951.891212] env[61728]: DEBUG oslo_concurrency.lockutils [req-f7857d60-43da-4db6-8f59-9756d4229107 req-3ec051fb-ee67-45a6-9ed0-cacae495e546 service nova] Acquiring lock "refresh_cache-0af88a53-bbe7-443e-a7dc-734e483bb730" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.891372] env[61728]: DEBUG oslo_concurrency.lockutils [req-f7857d60-43da-4db6-8f59-9756d4229107 req-3ec051fb-ee67-45a6-9ed0-cacae495e546 service nova] Acquired lock "refresh_cache-0af88a53-bbe7-443e-a7dc-734e483bb730" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.891562] env[61728]: DEBUG nova.network.neutron [req-f7857d60-43da-4db6-8f59-9756d4229107 req-3ec051fb-ee67-45a6-9ed0-cacae495e546 service nova] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Refreshing network info cache for port 814bcd52-6c0c-4efd-ae00-1cfb3375cd63 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 951.953533] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "3415687e-9b8d-42f8-9b18-93951b8ec7da" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.953901] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "3415687e-9b8d-42f8-9b18-93951b8ec7da" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.954197] env[61728]: DEBUG nova.compute.manager [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Going to confirm migration 5 {{(pid=61728) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 952.018958] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464546, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.176202] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464547, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.381111] env[61728]: DEBUG nova.scheduler.client.report [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.517156] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464546, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.834897} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.517472] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] a70d1c37-e792-4168-b7e1-9418b8cb7818/a70d1c37-e792-4168-b7e1-9418b8cb7818.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 952.517660] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 952.517930] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9326d961-3792-4d34-ad29-9392496fa9a4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.525173] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 952.525173] env[61728]: value = "task-464549" [ 952.525173] env[61728]: _type = "Task" [ 952.525173] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.533646] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464549, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.563898] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.564111] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.564342] env[61728]: DEBUG nova.network.neutron [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 952.564556] env[61728]: DEBUG nova.objects.instance [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lazy-loading 'info_cache' on Instance uuid 3415687e-9b8d-42f8-9b18-93951b8ec7da {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.658877] env[61728]: DEBUG nova.network.neutron [req-f7857d60-43da-4db6-8f59-9756d4229107 req-3ec051fb-ee67-45a6-9ed0-cacae495e546 service nova] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Updated VIF entry in instance network info cache for port 814bcd52-6c0c-4efd-ae00-1cfb3375cd63. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 952.659184] env[61728]: DEBUG nova.network.neutron [req-f7857d60-43da-4db6-8f59-9756d4229107 req-3ec051fb-ee67-45a6-9ed0-cacae495e546 service nova] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Updating instance_info_cache with network_info: [{"id": "814bcd52-6c0c-4efd-ae00-1cfb3375cd63", "address": "fa:16:3e:6f:a5:67", "network": {"id": "ed539cc4-acd5-4404-a241-e50f61a1d8e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2088844048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c78292171414530a1caba3dfd0f620f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap814bcd52-6c", "ovs_interfaceid": "814bcd52-6c0c-4efd-ae00-1cfb3375cd63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.676995] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464547, 'name': CreateVM_Task, 'duration_secs': 0.711746} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.677213] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 952.677897] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.678082] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.678452] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 952.678740] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab5f6665-bd25-40ae-a0d9-abebb7251c08 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.684425] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 952.684425] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524d081b-e84b-c060-cf5b-9d21ea8976a7" [ 952.684425] env[61728]: _type = "Task" [ 952.684425] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.695923] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524d081b-e84b-c060-cf5b-9d21ea8976a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.886854] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.424s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.888991] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.618s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.890633] env[61728]: INFO nova.compute.claims [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 952.920474] env[61728]: INFO nova.network.neutron [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Updating port 124d4780-15ec-4f17-a2c4-edf1517ed589 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 953.035172] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464549, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075357} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.035743] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 953.036745] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb4afe4-116b-4b96-9594-07fe0bb4be0e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.059842] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] a70d1c37-e792-4168-b7e1-9418b8cb7818/a70d1c37-e792-4168-b7e1-9418b8cb7818.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 953.060378] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bec508df-3b7e-40b4-8db9-1659e3d426c3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.083730] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 953.083730] env[61728]: value = "task-464550" [ 953.083730] env[61728]: _type = "Task" [ 953.083730] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.092064] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464550, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.162188] env[61728]: DEBUG oslo_concurrency.lockutils [req-f7857d60-43da-4db6-8f59-9756d4229107 req-3ec051fb-ee67-45a6-9ed0-cacae495e546 service nova] Releasing lock "refresh_cache-0af88a53-bbe7-443e-a7dc-734e483bb730" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.197358] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524d081b-e84b-c060-cf5b-9d21ea8976a7, 'name': SearchDatastore_Task, 'duration_secs': 0.020547} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.197721] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.198017] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.198362] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.198574] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.200041] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.200041] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-35caa144-779c-4489-aeba-25657aa5e6dd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.208296] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.209099] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 953.209259] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd388496-28e2-455b-9c90-c5a590cad25a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.214756] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 953.214756] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524a6963-80f0-e096-273f-fdddd86da72f" [ 953.214756] env[61728]: _type = "Task" [ 953.214756] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.223143] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524a6963-80f0-e096-273f-fdddd86da72f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.593628] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464550, 'name': ReconfigVM_Task, 'duration_secs': 0.297641} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.593957] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Reconfigured VM instance instance-0000005a to attach disk [datastore1] a70d1c37-e792-4168-b7e1-9418b8cb7818/a70d1c37-e792-4168-b7e1-9418b8cb7818.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 953.594524] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-66023a6c-8601-4e1f-9ba0-044da6a250c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.600484] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 953.600484] env[61728]: value = "task-464551" [ 953.600484] env[61728]: _type = "Task" [ 953.600484] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.608123] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464551, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.724736] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524a6963-80f0-e096-273f-fdddd86da72f, 'name': SearchDatastore_Task, 'duration_secs': 0.016757} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.725545] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce9df273-fd03-4593-a7de-eede356d2f27 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.730801] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 953.730801] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dd81d6-344d-f024-e01d-944d8329a0ca" [ 953.730801] env[61728]: _type = "Task" [ 953.730801] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.739140] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dd81d6-344d-f024-e01d-944d8329a0ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.774724] env[61728]: DEBUG nova.network.neutron [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating instance_info_cache with network_info: [{"id": "fbf3382b-a343-4cc4-9233-ebf4ac3139a4", "address": "fa:16:3e:e3:49:f4", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbf3382b-a3", "ovs_interfaceid": "fbf3382b-a343-4cc4-9233-ebf4ac3139a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.110451] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464551, 'name': Rename_Task, 'duration_secs': 0.142655} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.112670] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 954.113092] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8889b6e9-10cf-493f-8274-366b2406f80c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.120270] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 954.120270] env[61728]: value = "task-464552" [ 954.120270] env[61728]: _type = "Task" [ 954.120270] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.129734] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464552, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.131205] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b3c1dd-f06d-46f5-8bef-596eb67310aa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.137833] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09bf1f4-146f-4944-9904-5b6fb5079097 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.170187] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d1ea99-bea9-478d-9316-4e7f0a3e63a7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.178256] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08c1776-f834-40f6-b372-7dd0ed514521 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.191936] env[61728]: DEBUG nova.compute.provider_tree [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.241743] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dd81d6-344d-f024-e01d-944d8329a0ca, 'name': SearchDatastore_Task, 'duration_secs': 0.012712} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.242049] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.242323] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 0af88a53-bbe7-443e-a7dc-734e483bb730/0af88a53-bbe7-443e-a7dc-734e483bb730.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 954.242586] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cf1cf21d-b61c-4eb5-87d1-4031cc2add44 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.249020] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 954.249020] env[61728]: value = "task-464553" [ 954.249020] env[61728]: _type = "Task" [ 954.249020] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.256739] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464553, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.277051] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "refresh_cache-3415687e-9b8d-42f8-9b18-93951b8ec7da" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.277351] env[61728]: DEBUG nova.objects.instance [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lazy-loading 'migration_context' on Instance uuid 3415687e-9b8d-42f8-9b18-93951b8ec7da {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.303883] env[61728]: DEBUG nova.compute.manager [req-620039fb-ba06-4ddd-8038-b3da128f2c74 req-4b4ffa9f-ac41-409b-bb6d-ebc65b433d12 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Received event network-vif-plugged-124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 954.304113] env[61728]: DEBUG oslo_concurrency.lockutils [req-620039fb-ba06-4ddd-8038-b3da128f2c74 req-4b4ffa9f-ac41-409b-bb6d-ebc65b433d12 service nova] Acquiring lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.304321] env[61728]: DEBUG oslo_concurrency.lockutils [req-620039fb-ba06-4ddd-8038-b3da128f2c74 req-4b4ffa9f-ac41-409b-bb6d-ebc65b433d12 service nova] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.304486] env[61728]: DEBUG oslo_concurrency.lockutils [req-620039fb-ba06-4ddd-8038-b3da128f2c74 req-4b4ffa9f-ac41-409b-bb6d-ebc65b433d12 service nova] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.304654] env[61728]: DEBUG nova.compute.manager [req-620039fb-ba06-4ddd-8038-b3da128f2c74 req-4b4ffa9f-ac41-409b-bb6d-ebc65b433d12 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] No waiting events found dispatching network-vif-plugged-124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 954.304833] env[61728]: WARNING nova.compute.manager [req-620039fb-ba06-4ddd-8038-b3da128f2c74 req-4b4ffa9f-ac41-409b-bb6d-ebc65b433d12 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Received unexpected event network-vif-plugged-124d4780-15ec-4f17-a2c4-edf1517ed589 for instance with vm_state shelved_offloaded and task_state spawning. [ 954.386604] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.386604] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquired lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.386604] env[61728]: DEBUG nova.network.neutron [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 954.437852] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 954.438167] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122162', 'volume_id': 'fe562e41-06ae-47c9-9eda-4f075c89cd34', 'name': 'volume-fe562e41-06ae-47c9-9eda-4f075c89cd34', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f9f07464-9f43-43fd-8895-2673861747fa', 'attached_at': '', 'detached_at': '', 'volume_id': 'fe562e41-06ae-47c9-9eda-4f075c89cd34', 'serial': 'fe562e41-06ae-47c9-9eda-4f075c89cd34'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 954.439085] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ed5985-88ba-4e61-85a4-153573ddba1b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.459116] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d7a2787-3c5d-4d51-9f3c-6b919d288e8c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.484747] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] volume-fe562e41-06ae-47c9-9eda-4f075c89cd34/volume-fe562e41-06ae-47c9-9eda-4f075c89cd34.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.485234] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cfeee1dc-55e4-413d-8e7b-91ae54b7a64a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.505937] env[61728]: DEBUG oslo_vmware.api [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 954.505937] env[61728]: value = "task-464554" [ 954.505937] env[61728]: _type = "Task" [ 954.505937] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.516303] env[61728]: DEBUG oslo_vmware.api [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464554, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.632540] env[61728]: DEBUG oslo_vmware.api [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464552, 'name': PowerOnVM_Task, 'duration_secs': 0.467613} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.632965] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 954.633049] env[61728]: INFO nova.compute.manager [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Took 7.65 seconds to spawn the instance on the hypervisor. [ 954.633241] env[61728]: DEBUG nova.compute.manager [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.634126] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9819d15-2208-4d46-9585-44ea4c6cf6f8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.695135] env[61728]: DEBUG nova.scheduler.client.report [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.759235] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464553, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470633} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.759477] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 0af88a53-bbe7-443e-a7dc-734e483bb730/0af88a53-bbe7-443e-a7dc-734e483bb730.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 954.759690] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 954.759939] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3b8326df-6bf1-4334-8059-916b6e4c7d47 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.765969] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 954.765969] env[61728]: value = "task-464555" [ 954.765969] env[61728]: _type = "Task" [ 954.765969] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.774365] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464555, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.780296] env[61728]: DEBUG nova.objects.base [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Object Instance<3415687e-9b8d-42f8-9b18-93951b8ec7da> lazy-loaded attributes: info_cache,migration_context {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 954.781509] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30c47e0-77d1-4fd9-bbe2-02e799c5504d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.799564] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b275df1-d45b-4173-84ff-39aaf4da7e14 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.804844] env[61728]: DEBUG oslo_vmware.api [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 954.804844] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dc5e51-b51d-903b-9af3-6347c820cce2" [ 954.804844] env[61728]: _type = "Task" [ 954.804844] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.812813] env[61728]: DEBUG oslo_vmware.api [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dc5e51-b51d-903b-9af3-6347c820cce2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.016955] env[61728]: DEBUG oslo_vmware.api [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464554, 'name': ReconfigVM_Task, 'duration_secs': 0.488484} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.017291] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Reconfigured VM instance instance-00000051 to attach disk [datastore1] volume-fe562e41-06ae-47c9-9eda-4f075c89cd34/volume-fe562e41-06ae-47c9-9eda-4f075c89cd34.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.024819] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb273eab-a829-41a9-853a-17091f36b15f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.041320] env[61728]: DEBUG oslo_vmware.api [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 955.041320] env[61728]: value = "task-464556" [ 955.041320] env[61728]: _type = "Task" [ 955.041320] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.049782] env[61728]: DEBUG oslo_vmware.api [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464556, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.105283] env[61728]: DEBUG nova.network.neutron [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Updating instance_info_cache with network_info: [{"id": "124d4780-15ec-4f17-a2c4-edf1517ed589", "address": "fa:16:3e:6d:47:93", "network": {"id": "6a1908cc-f6aa-4dd7-970e-c8a07e1b16ee", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2047209328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "01839050e56e4dbfa6e22471a3c60998", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap124d4780-15", "ovs_interfaceid": "124d4780-15ec-4f17-a2c4-edf1517ed589", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.155724] env[61728]: INFO nova.compute.manager [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Took 45.33 seconds to build instance. [ 955.200174] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.311s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.200594] env[61728]: DEBUG nova.compute.manager [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 955.203360] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.165s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.203587] env[61728]: DEBUG nova.objects.instance [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lazy-loading 'resources' on Instance uuid ae25dc25-f600-471f-aaaa-f166ed691bf3 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.275936] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464555, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058808} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.276219] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.276955] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57587573-6c0d-4857-b6f5-3215c5a6dc24 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.297952] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 0af88a53-bbe7-443e-a7dc-734e483bb730/0af88a53-bbe7-443e-a7dc-734e483bb730.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.298085] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8bc621a2-a9d3-41f0-8461-e1145b608f49 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.324755] env[61728]: DEBUG oslo_vmware.api [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dc5e51-b51d-903b-9af3-6347c820cce2, 'name': SearchDatastore_Task, 'duration_secs': 0.006687} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.326139] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.326396] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 955.326396] env[61728]: value = "task-464557" [ 955.326396] env[61728]: _type = "Task" [ 955.326396] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.334336] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464557, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.550926] env[61728]: DEBUG oslo_vmware.api [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464556, 'name': ReconfigVM_Task, 'duration_secs': 0.155108} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.551272] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122162', 'volume_id': 'fe562e41-06ae-47c9-9eda-4f075c89cd34', 'name': 'volume-fe562e41-06ae-47c9-9eda-4f075c89cd34', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f9f07464-9f43-43fd-8895-2673861747fa', 'attached_at': '', 'detached_at': '', 'volume_id': 'fe562e41-06ae-47c9-9eda-4f075c89cd34', 'serial': 'fe562e41-06ae-47c9-9eda-4f075c89cd34'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 955.608307] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Releasing lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.631479] env[61728]: DEBUG nova.virt.hardware [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='1a2be8afa0d69d727b451916526d2679',container_format='bare',created_at=2024-08-29T12:28:48Z,direct_url=,disk_format='vmdk',id=6bf918f6-e123-44bd-a072-fe70029d1786,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1687284999-shelved',owner='01839050e56e4dbfa6e22471a3c60998',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-08-29T12:29:03Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 955.631817] env[61728]: DEBUG nova.virt.hardware [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 955.632025] env[61728]: DEBUG nova.virt.hardware [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 955.632234] env[61728]: DEBUG nova.virt.hardware [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 955.632392] env[61728]: DEBUG nova.virt.hardware [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 955.632550] env[61728]: DEBUG nova.virt.hardware [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 955.632790] env[61728]: DEBUG nova.virt.hardware [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 955.633051] env[61728]: DEBUG nova.virt.hardware [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 955.633230] env[61728]: DEBUG nova.virt.hardware [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 955.633409] env[61728]: DEBUG nova.virt.hardware [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 955.633591] env[61728]: DEBUG nova.virt.hardware [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 955.634470] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f98cc85-654a-4528-ad3c-77b33947acf9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.644039] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d458408f-6db9-453e-af76-a0fa43e54f54 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.657168] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:47:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48512b02-ad5c-4105-ba7d-fd4775acf8e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '124d4780-15ec-4f17-a2c4-edf1517ed589', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 955.664568] env[61728]: DEBUG oslo.service.loopingcall [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.665016] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ec6dc8bd-9a9a-4a04-b933-3eba1dd911b0 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "a70d1c37-e792-4168-b7e1-9418b8cb7818" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.849s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.665262] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 955.665482] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea19f131-6424-4a3a-9d5e-b068f6652b51 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.684050] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 955.684050] env[61728]: value = "task-464558" [ 955.684050] env[61728]: _type = "Task" [ 955.684050] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.691452] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464558, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.706616] env[61728]: DEBUG nova.compute.utils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 955.708014] env[61728]: DEBUG nova.compute.manager [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 955.708199] env[61728]: DEBUG nova.network.neutron [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 955.754485] env[61728]: DEBUG nova.policy [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e35028155c743e4bcf2ab91be37b201', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c035f75f40864559b47b643fcfca56e1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 955.838694] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464557, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.034955] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43999f71-0306-4790-b2a1-99324c6b692a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.042994] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcfed825-eeca-4506-9602-e044078259a4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.084631] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93f3719-d257-41e7-a4e9-d3f44960e005 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.092616] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6c0aa96-524a-4f7d-bded-3538c4daba54 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.112413] env[61728]: DEBUG nova.compute.provider_tree [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.117275] env[61728]: DEBUG nova.network.neutron [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Successfully created port: 67077c45-d87f-4d2d-a48b-d23a96d35f8a {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 956.197208] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464558, 'name': CreateVM_Task, 'duration_secs': 0.421344} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.197208] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 956.198624] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6bf918f6-e123-44bd-a072-fe70029d1786" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.198891] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6bf918f6-e123-44bd-a072-fe70029d1786" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.199683] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6bf918f6-e123-44bd-a072-fe70029d1786" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 956.199805] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7194e3ec-d9de-45cf-9d1b-ebfa1df6e605 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.204926] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 956.204926] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52970443-632b-5706-fe31-7501e95ae616" [ 956.204926] env[61728]: _type = "Task" [ 956.204926] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.214862] env[61728]: DEBUG nova.compute.manager [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 956.217984] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52970443-632b-5706-fe31-7501e95ae616, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.337434] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464557, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.461408] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "705c8838-f605-46fa-b036-ec212f673f66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.461658] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "705c8838-f605-46fa-b036-ec212f673f66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.491229] env[61728]: DEBUG nova.compute.manager [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Received event network-changed-124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 956.491500] env[61728]: DEBUG nova.compute.manager [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Refreshing instance network info cache due to event network-changed-124d4780-15ec-4f17-a2c4-edf1517ed589. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 956.491627] env[61728]: DEBUG oslo_concurrency.lockutils [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] Acquiring lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.491805] env[61728]: DEBUG oslo_concurrency.lockutils [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] Acquired lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.491978] env[61728]: DEBUG nova.network.neutron [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Refreshing network info cache for port 124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 956.618341] env[61728]: DEBUG nova.scheduler.client.report [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.644779] env[61728]: DEBUG nova.objects.instance [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'flavor' on Instance uuid f9f07464-9f43-43fd-8895-2673861747fa {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.715931] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6bf918f6-e123-44bd-a072-fe70029d1786" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.716605] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Processing image 6bf918f6-e123-44bd-a072-fe70029d1786 {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 956.716605] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6bf918f6-e123-44bd-a072-fe70029d1786/6bf918f6-e123-44bd-a072-fe70029d1786.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.716904] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6bf918f6-e123-44bd-a072-fe70029d1786/6bf918f6-e123-44bd-a072-fe70029d1786.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.716904] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 956.717070] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4cb32a63-5c4a-4638-8ee3-9421ffc27338 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.731320] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 956.731528] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 956.732329] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30a5ab6b-bc70-43ec-8264-39447de93aa5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.738193] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 956.738193] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ce59ba-2195-2bed-ce93-98f2045775ef" [ 956.738193] env[61728]: _type = "Task" [ 956.738193] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.746641] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ce59ba-2195-2bed-ce93-98f2045775ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.842316] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464557, 'name': ReconfigVM_Task, 'duration_secs': 1.132426} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.842659] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 0af88a53-bbe7-443e-a7dc-734e483bb730/0af88a53-bbe7-443e-a7dc-734e483bb730.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.843341] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0acfbc97-7465-4dd1-893d-64b40fd01407 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.850324] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 956.850324] env[61728]: value = "task-464559" [ 956.850324] env[61728]: _type = "Task" [ 956.850324] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.865018] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464559, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.964652] env[61728]: DEBUG nova.compute.manager [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 957.126159] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.921s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.127929] env[61728]: DEBUG oslo_concurrency.lockutils [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.988s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.128221] env[61728]: DEBUG nova.objects.instance [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Lazy-loading 'resources' on Instance uuid 2400abb1-8a07-4f6b-8818-778a2105beb8 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.148784] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4d1a851b-ec8d-445f-9826-8153de68a538 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "f9f07464-9f43-43fd-8895-2673861747fa" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.338s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.150158] env[61728]: INFO nova.scheduler.client.report [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleted allocations for instance ae25dc25-f600-471f-aaaa-f166ed691bf3 [ 957.218136] env[61728]: DEBUG nova.network.neutron [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Updated VIF entry in instance network info cache for port 124d4780-15ec-4f17-a2c4-edf1517ed589. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 957.218598] env[61728]: DEBUG nova.network.neutron [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Updating instance_info_cache with network_info: [{"id": "124d4780-15ec-4f17-a2c4-edf1517ed589", "address": "fa:16:3e:6d:47:93", "network": {"id": "6a1908cc-f6aa-4dd7-970e-c8a07e1b16ee", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2047209328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "01839050e56e4dbfa6e22471a3c60998", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap124d4780-15", "ovs_interfaceid": "124d4780-15ec-4f17-a2c4-edf1517ed589", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.226620] env[61728]: DEBUG nova.compute.manager [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 957.251445] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Preparing fetch location {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 957.251883] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Fetch image to [datastore1] OSTACK_IMG_85e605de-cdc3-4762-80b1-95d21bd5f31b/OSTACK_IMG_85e605de-cdc3-4762-80b1-95d21bd5f31b.vmdk {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 957.252344] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Downloading stream optimized image 6bf918f6-e123-44bd-a072-fe70029d1786 to [datastore1] OSTACK_IMG_85e605de-cdc3-4762-80b1-95d21bd5f31b/OSTACK_IMG_85e605de-cdc3-4762-80b1-95d21bd5f31b.vmdk on the data store datastore1 as vApp {{(pid=61728) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 957.252679] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Downloading image file data 6bf918f6-e123-44bd-a072-fe70029d1786 to the ESX as VM named 'OSTACK_IMG_85e605de-cdc3-4762-80b1-95d21bd5f31b' {{(pid=61728) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 957.262095] env[61728]: DEBUG nova.virt.hardware [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 957.262842] env[61728]: DEBUG nova.virt.hardware [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 957.263143] env[61728]: DEBUG nova.virt.hardware [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 957.263630] env[61728]: DEBUG nova.virt.hardware [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 957.263884] env[61728]: DEBUG nova.virt.hardware [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 957.264214] env[61728]: DEBUG nova.virt.hardware [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 957.264587] env[61728]: DEBUG nova.virt.hardware [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 957.264870] env[61728]: DEBUG nova.virt.hardware [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 957.265177] env[61728]: DEBUG nova.virt.hardware [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 957.265466] env[61728]: DEBUG nova.virt.hardware [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 957.265843] env[61728]: DEBUG nova.virt.hardware [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 957.266926] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59724cdb-64d4-4e6b-a55d-f91166e97c71 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.275850] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6328d86-c6db-4423-ada0-10d1e1cd21d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.352386] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 957.352386] env[61728]: value = "resgroup-9" [ 957.352386] env[61728]: _type = "ResourcePool" [ 957.352386] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 957.353079] env[61728]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-81c9f743-d30d-41ab-aa11-2115f8639d40 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.381137] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464559, 'name': Rename_Task, 'duration_secs': 0.206286} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.382561] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 957.382915] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lease: (returnval){ [ 957.382915] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb6822-8eef-7c0a-154d-7fc662c7ff22" [ 957.382915] env[61728]: _type = "HttpNfcLease" [ 957.382915] env[61728]: } obtained for vApp import into resource pool (val){ [ 957.382915] env[61728]: value = "resgroup-9" [ 957.382915] env[61728]: _type = "ResourcePool" [ 957.382915] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 957.383159] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the lease: (returnval){ [ 957.383159] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb6822-8eef-7c0a-154d-7fc662c7ff22" [ 957.383159] env[61728]: _type = "HttpNfcLease" [ 957.383159] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 957.383361] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8676c040-874c-4e9d-a938-ec383b40f044 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.394702] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 957.394702] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb6822-8eef-7c0a-154d-7fc662c7ff22" [ 957.394702] env[61728]: _type = "HttpNfcLease" [ 957.394702] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 957.396358] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 957.396358] env[61728]: value = "task-464561" [ 957.396358] env[61728]: _type = "Task" [ 957.396358] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.405768] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464561, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.490464] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.641155] env[61728]: DEBUG nova.network.neutron [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Successfully updated port: 67077c45-d87f-4d2d-a48b-d23a96d35f8a {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 957.658701] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3c46d77-b287-4a17-be36-6da94e7a9363 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "ae25dc25-f600-471f-aaaa-f166ed691bf3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.075s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.716934] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.717366] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.722575] env[61728]: DEBUG oslo_concurrency.lockutils [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] Releasing lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.722975] env[61728]: DEBUG nova.compute.manager [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Received event network-changed-c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 957.723325] env[61728]: DEBUG nova.compute.manager [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Refreshing instance network info cache due to event network-changed-c8407eac-4337-4ae3-a071-08f48bf87248. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 957.723677] env[61728]: DEBUG oslo_concurrency.lockutils [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] Acquiring lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.723921] env[61728]: DEBUG oslo_concurrency.lockutils [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] Acquired lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.724243] env[61728]: DEBUG nova.network.neutron [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Refreshing network info cache for port c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 957.893220] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 957.893220] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb6822-8eef-7c0a-154d-7fc662c7ff22" [ 957.893220] env[61728]: _type = "HttpNfcLease" [ 957.893220] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 957.906240] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464561, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.957947] env[61728]: DEBUG nova.compute.manager [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Stashing vm_state: active {{(pid=61728) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 958.010658] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c71110f-1b3f-41d7-9641-b2ecd85b20c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.019557] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea64eb7d-a5f1-4a2a-b691-961b3a8276ab {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.052770] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca310843-382b-47d6-82ee-a2db8b8c37cf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.060432] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c9f1cc-fcb6-4ab9-9197-3dd1a955fa9e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.076607] env[61728]: DEBUG nova.compute.provider_tree [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.143318] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.143489] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.143636] env[61728]: DEBUG nova.network.neutron [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 958.221735] env[61728]: DEBUG nova.compute.manager [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 958.393680] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 958.393680] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb6822-8eef-7c0a-154d-7fc662c7ff22" [ 958.393680] env[61728]: _type = "HttpNfcLease" [ 958.393680] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 958.405801] env[61728]: DEBUG oslo_vmware.api [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464561, 'name': PowerOnVM_Task, 'duration_secs': 0.659963} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.409386] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 958.409756] env[61728]: INFO nova.compute.manager [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Took 8.91 seconds to spawn the instance on the hypervisor. [ 958.410579] env[61728]: DEBUG nova.compute.manager [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.413293] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0af5ba-4aaf-4393-a299-a417fcbdc1b2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.450437] env[61728]: DEBUG nova.network.neutron [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updated VIF entry in instance network info cache for port c8407eac-4337-4ae3-a071-08f48bf87248. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 958.450720] env[61728]: DEBUG nova.network.neutron [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updating instance_info_cache with network_info: [{"id": "c8407eac-4337-4ae3-a071-08f48bf87248", "address": "fa:16:3e:2d:a6:d6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8407eac-43", "ovs_interfaceid": "c8407eac-4337-4ae3-a071-08f48bf87248", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.482823] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.535030] env[61728]: DEBUG nova.compute.manager [req-d4046f74-0032-493b-8f4a-242c708c3d86 req-91c641c7-beb8-4355-8a11-459125146236 service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Received event network-vif-plugged-67077c45-d87f-4d2d-a48b-d23a96d35f8a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 958.535343] env[61728]: DEBUG oslo_concurrency.lockutils [req-d4046f74-0032-493b-8f4a-242c708c3d86 req-91c641c7-beb8-4355-8a11-459125146236 service nova] Acquiring lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.535427] env[61728]: DEBUG oslo_concurrency.lockutils [req-d4046f74-0032-493b-8f4a-242c708c3d86 req-91c641c7-beb8-4355-8a11-459125146236 service nova] Lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.535547] env[61728]: DEBUG oslo_concurrency.lockutils [req-d4046f74-0032-493b-8f4a-242c708c3d86 req-91c641c7-beb8-4355-8a11-459125146236 service nova] Lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.535727] env[61728]: DEBUG nova.compute.manager [req-d4046f74-0032-493b-8f4a-242c708c3d86 req-91c641c7-beb8-4355-8a11-459125146236 service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] No waiting events found dispatching network-vif-plugged-67077c45-d87f-4d2d-a48b-d23a96d35f8a {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 958.535899] env[61728]: WARNING nova.compute.manager [req-d4046f74-0032-493b-8f4a-242c708c3d86 req-91c641c7-beb8-4355-8a11-459125146236 service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Received unexpected event network-vif-plugged-67077c45-d87f-4d2d-a48b-d23a96d35f8a for instance with vm_state building and task_state spawning. [ 958.536075] env[61728]: DEBUG nova.compute.manager [req-d4046f74-0032-493b-8f4a-242c708c3d86 req-91c641c7-beb8-4355-8a11-459125146236 service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Received event network-changed-67077c45-d87f-4d2d-a48b-d23a96d35f8a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 958.536239] env[61728]: DEBUG nova.compute.manager [req-d4046f74-0032-493b-8f4a-242c708c3d86 req-91c641c7-beb8-4355-8a11-459125146236 service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Refreshing instance network info cache due to event network-changed-67077c45-d87f-4d2d-a48b-d23a96d35f8a. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 958.536414] env[61728]: DEBUG oslo_concurrency.lockutils [req-d4046f74-0032-493b-8f4a-242c708c3d86 req-91c641c7-beb8-4355-8a11-459125146236 service nova] Acquiring lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.565407] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "172129ff-136e-4855-8659-3f1ac88dcbd2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.565626] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "172129ff-136e-4855-8659-3f1ac88dcbd2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.580719] env[61728]: DEBUG nova.scheduler.client.report [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.674320] env[61728]: DEBUG nova.network.neutron [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 958.742406] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.792825] env[61728]: DEBUG nova.network.neutron [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Updating instance_info_cache with network_info: [{"id": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "address": "fa:16:3e:d1:31:88", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67077c45-d8", "ovs_interfaceid": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.895293] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 958.895293] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb6822-8eef-7c0a-154d-7fc662c7ff22" [ 958.895293] env[61728]: _type = "HttpNfcLease" [ 958.895293] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 958.929788] env[61728]: INFO nova.compute.manager [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Took 42.78 seconds to build instance. [ 958.953891] env[61728]: DEBUG oslo_concurrency.lockutils [req-facaafa3-2690-4535-ad07-baebb120a713 req-f62b7c52-d234-467e-b054-98922d2ffbd5 service nova] Releasing lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.068051] env[61728]: DEBUG nova.compute.manager [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 959.084756] env[61728]: DEBUG oslo_concurrency.lockutils [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.957s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.087101] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.925s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.088822] env[61728]: INFO nova.compute.claims [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 959.106079] env[61728]: INFO nova.scheduler.client.report [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Deleted allocations for instance 2400abb1-8a07-4f6b-8818-778a2105beb8 [ 959.296657] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.297068] env[61728]: DEBUG nova.compute.manager [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Instance network_info: |[{"id": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "address": "fa:16:3e:d1:31:88", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67077c45-d8", "ovs_interfaceid": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 959.297431] env[61728]: DEBUG oslo_concurrency.lockutils [req-d4046f74-0032-493b-8f4a-242c708c3d86 req-91c641c7-beb8-4355-8a11-459125146236 service nova] Acquired lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.297652] env[61728]: DEBUG nova.network.neutron [req-d4046f74-0032-493b-8f4a-242c708c3d86 req-91c641c7-beb8-4355-8a11-459125146236 service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Refreshing network info cache for port 67077c45-d87f-4d2d-a48b-d23a96d35f8a {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 959.298897] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:31:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2be3fdb5-359e-43bd-8c20-2ff00e81db55', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '67077c45-d87f-4d2d-a48b-d23a96d35f8a', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 959.307015] env[61728]: DEBUG oslo.service.loopingcall [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.307682] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 959.307949] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-735d0389-ab53-4538-b39f-a013177f0704 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.329312] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 959.329312] env[61728]: value = "task-464562" [ 959.329312] env[61728]: _type = "Task" [ 959.329312] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.337831] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464562, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.397467] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.397467] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb6822-8eef-7c0a-154d-7fc662c7ff22" [ 959.397467] env[61728]: _type = "HttpNfcLease" [ 959.397467] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 959.432355] env[61728]: DEBUG oslo_concurrency.lockutils [None req-abfc01c2-7fb0-4d17-ae63-bfe10c73547f tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "0af88a53-bbe7-443e-a7dc-734e483bb730" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.291s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.586217] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.614121] env[61728]: DEBUG oslo_concurrency.lockutils [None req-04d88740-db53-4dfd-8514-8bdb818b0c0e tempest-ServerMetadataTestJSON-263854492 tempest-ServerMetadataTestJSON-263854492-project-member] Lock "2400abb1-8a07-4f6b-8818-778a2105beb8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.329s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.840027] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464562, 'name': CreateVM_Task, 'duration_secs': 0.345225} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.840027] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 959.840447] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.840622] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.840997] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.841221] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ba7bec6-ae2a-4619-988f-63fb20e6983e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.845647] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 959.845647] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524ea4e0-2bc0-9577-16b3-bc1915797aef" [ 959.845647] env[61728]: _type = "Task" [ 959.845647] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.853606] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524ea4e0-2bc0-9577-16b3-bc1915797aef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.896478] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.896478] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb6822-8eef-7c0a-154d-7fc662c7ff22" [ 959.896478] env[61728]: _type = "HttpNfcLease" [ 959.896478] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 960.203273] env[61728]: DEBUG nova.network.neutron [req-d4046f74-0032-493b-8f4a-242c708c3d86 req-91c641c7-beb8-4355-8a11-459125146236 service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Updated VIF entry in instance network info cache for port 67077c45-d87f-4d2d-a48b-d23a96d35f8a. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 960.203759] env[61728]: DEBUG nova.network.neutron [req-d4046f74-0032-493b-8f4a-242c708c3d86 req-91c641c7-beb8-4355-8a11-459125146236 service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Updating instance_info_cache with network_info: [{"id": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "address": "fa:16:3e:d1:31:88", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67077c45-d8", "ovs_interfaceid": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.363904] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524ea4e0-2bc0-9577-16b3-bc1915797aef, 'name': SearchDatastore_Task, 'duration_secs': 0.0107} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.364288] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.364534] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 960.364774] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.364931] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.366206] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 960.366495] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74fa3a73-0549-4551-8df7-520927327f51 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.378020] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 960.378020] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 960.378020] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbf6bbb0-3146-4295-8935-86b300b807c9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.386082] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 960.386082] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5285a35d-8a4e-d41f-7d00-b06a8b63f747" [ 960.386082] env[61728]: _type = "Task" [ 960.386082] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.404457] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5285a35d-8a4e-d41f-7d00-b06a8b63f747, 'name': SearchDatastore_Task, 'duration_secs': 0.008847} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.405612] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1350cab-fbf5-4076-ad2d-18bcc6e263a8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.408852] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 960.408852] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb6822-8eef-7c0a-154d-7fc662c7ff22" [ 960.408852] env[61728]: _type = "HttpNfcLease" [ 960.408852] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 960.413424] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 960.413424] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f7c9f5-058c-25c6-2a89-41e32b358a14" [ 960.413424] env[61728]: _type = "Task" [ 960.413424] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.421201] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f7c9f5-058c-25c6-2a89-41e32b358a14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.427888] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf7eadc-6945-475d-a89e-7eb99e8fd859 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.438301] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2498ce4e-5fad-468b-89fc-46607a33cc82 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.472750] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5597e36-5090-4138-ad3b-aefd532b77d8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.482882] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa9f3e2-5a9c-4b0d-8a88-e08213fe62a5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.497485] env[61728]: DEBUG nova.compute.provider_tree [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.706753] env[61728]: DEBUG oslo_concurrency.lockutils [req-d4046f74-0032-493b-8f4a-242c708c3d86 req-91c641c7-beb8-4355-8a11-459125146236 service nova] Releasing lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.901940] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 960.901940] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb6822-8eef-7c0a-154d-7fc662c7ff22" [ 960.901940] env[61728]: _type = "HttpNfcLease" [ 960.901940] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 960.926619] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f7c9f5-058c-25c6-2a89-41e32b358a14, 'name': SearchDatastore_Task, 'duration_secs': 0.009736} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.926941] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.927292] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a/4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 960.927584] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-16282b85-cfc1-408d-9e05-a5fb475d786e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.935829] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 960.935829] env[61728]: value = "task-464563" [ 960.935829] env[61728]: _type = "Task" [ 960.935829] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.944961] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464563, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.979908] env[61728]: DEBUG nova.compute.manager [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 960.980810] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1407212e-25f3-4f75-adc8-17beb899abde {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.004344] env[61728]: DEBUG nova.scheduler.client.report [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 961.403809] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 961.403809] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb6822-8eef-7c0a-154d-7fc662c7ff22" [ 961.403809] env[61728]: _type = "HttpNfcLease" [ 961.403809] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 961.449207] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464563, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.492095] env[61728]: INFO nova.compute.manager [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] instance snapshotting [ 961.495389] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fadfc2b-39b5-4271-b6a8-622b276339fb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.517032] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.517032] env[61728]: DEBUG nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 961.517430] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.324s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.518765] env[61728]: INFO nova.compute.claims [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 961.521754] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36852bde-3337-4e75-9066-737dcb2d8f16 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.900656] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 961.900656] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb6822-8eef-7c0a-154d-7fc662c7ff22" [ 961.900656] env[61728]: _type = "HttpNfcLease" [ 961.900656] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 961.945797] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464563, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.614119} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.946152] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a/4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 961.946468] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 961.946759] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b0989b3-a0a2-4b12-999d-1446e3aa3d17 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.953485] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 961.953485] env[61728]: value = "task-464564" [ 961.953485] env[61728]: _type = "Task" [ 961.953485] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.961203] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464564, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.027757] env[61728]: DEBUG nova.compute.utils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 962.031802] env[61728]: DEBUG nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 962.032333] env[61728]: DEBUG nova.network.neutron [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 962.035648] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 962.035648] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f071d198-434e-49c8-8056-7e32b3c76bbd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.045216] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 962.045216] env[61728]: value = "task-464565" [ 962.045216] env[61728]: _type = "Task" [ 962.045216] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.054720] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464565, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.086055] env[61728]: DEBUG nova.policy [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8938cfb5439247d6b95adfe040babded', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9523aa656ff2459b94bb28d133fa4b05', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 962.342699] env[61728]: DEBUG nova.network.neutron [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Successfully created port: 94c4dd39-2357-472e-8680-6d65f1755e09 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 962.401390] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 962.401390] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb6822-8eef-7c0a-154d-7fc662c7ff22" [ 962.401390] env[61728]: _type = "HttpNfcLease" [ 962.401390] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 962.401625] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 962.401625] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52eb6822-8eef-7c0a-154d-7fc662c7ff22" [ 962.401625] env[61728]: _type = "HttpNfcLease" [ 962.401625] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 962.402571] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66df442b-1ede-4973-b66f-d8b20e17b087 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.411356] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234d3f3-005b-2d61-2f04-3ee093b98279/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 962.411543] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234d3f3-005b-2d61-2f04-3ee093b98279/disk-0.vmdk. {{(pid=61728) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 962.480022] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3a20728f-197d-4cd7-a5bb-beaab2823073 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.480903] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464564, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094745} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.481206] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 962.482407] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c547fa-9b7b-4b54-b740-88bc7c45a4ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.505192] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a/4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.505496] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c372baa8-8c50-4c90-9eec-b2f9dc61e8c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.524943] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 962.524943] env[61728]: value = "task-464566" [ 962.524943] env[61728]: _type = "Task" [ 962.524943] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.533557] env[61728]: DEBUG nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 962.536147] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464566, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.554604] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464565, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.878872] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48884237-15f1-4f85-bde2-730d70af8d21 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.891456] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b65dda8-544c-4ceb-92f9-6a284e886ba8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.923780] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4637007a-3d10-4b18-8d8f-15a0345b61c3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.933564] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09de7104-9fe2-4798-a1eb-c83ddb05a789 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.953555] env[61728]: DEBUG nova.compute.provider_tree [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.035351] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464566, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.055834] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464565, 'name': CreateSnapshot_Task, 'duration_secs': 0.976035} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.057468] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 963.058173] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f8bc4b-0bbf-48f6-b122-8aedee4a5007 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.460037] env[61728]: DEBUG nova.scheduler.client.report [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.538097] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464566, 'name': ReconfigVM_Task, 'duration_secs': 0.586284} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.540071] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a/4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.541122] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-46a4e2ac-6a71-424a-8ab2-d02eb0b3240f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.543936] env[61728]: DEBUG nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 963.550339] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 963.550339] env[61728]: value = "task-464567" [ 963.550339] env[61728]: _type = "Task" [ 963.550339] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.560465] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464567, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.573777] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 963.574097] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 963.574400] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 963.574676] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 963.574859] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 963.575116] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 963.575372] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 963.575570] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 963.575764] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 963.575943] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 963.576136] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 963.583432] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 963.584277] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c3ba2e-9682-44db-bf8a-a452c01a8bc1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.590376] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-69f1276f-b28e-4bac-b550-767511508d03 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.602473] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc7bce3-ae6c-4f79-834c-8ef9d79b4b41 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.606704] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 963.606704] env[61728]: value = "task-464568" [ 963.606704] env[61728]: _type = "Task" [ 963.606704] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.628170] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464568, 'name': CloneVM_Task} progress is 12%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.766995] env[61728]: DEBUG nova.compute.manager [req-2501d2cd-af26-4437-b8df-90a58074bcb8 req-09cf607f-4462-46c2-afa5-9a977faef216 service nova] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Received event network-vif-plugged-94c4dd39-2357-472e-8680-6d65f1755e09 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 963.767292] env[61728]: DEBUG oslo_concurrency.lockutils [req-2501d2cd-af26-4437-b8df-90a58074bcb8 req-09cf607f-4462-46c2-afa5-9a977faef216 service nova] Acquiring lock "db547fea-4d44-440e-ac3e-29ae9b2c547b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.767621] env[61728]: DEBUG oslo_concurrency.lockutils [req-2501d2cd-af26-4437-b8df-90a58074bcb8 req-09cf607f-4462-46c2-afa5-9a977faef216 service nova] Lock "db547fea-4d44-440e-ac3e-29ae9b2c547b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.768100] env[61728]: DEBUG oslo_concurrency.lockutils [req-2501d2cd-af26-4437-b8df-90a58074bcb8 req-09cf607f-4462-46c2-afa5-9a977faef216 service nova] Lock "db547fea-4d44-440e-ac3e-29ae9b2c547b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.768313] env[61728]: DEBUG nova.compute.manager [req-2501d2cd-af26-4437-b8df-90a58074bcb8 req-09cf607f-4462-46c2-afa5-9a977faef216 service nova] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] No waiting events found dispatching network-vif-plugged-94c4dd39-2357-472e-8680-6d65f1755e09 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 963.768534] env[61728]: WARNING nova.compute.manager [req-2501d2cd-af26-4437-b8df-90a58074bcb8 req-09cf607f-4462-46c2-afa5-9a977faef216 service nova] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Received unexpected event network-vif-plugged-94c4dd39-2357-472e-8680-6d65f1755e09 for instance with vm_state building and task_state spawning. [ 963.859114] env[61728]: DEBUG nova.network.neutron [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Successfully updated port: 94c4dd39-2357-472e-8680-6d65f1755e09 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 963.964694] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.965622] env[61728]: DEBUG nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 963.970939] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.060s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.973273] env[61728]: INFO nova.compute.claims [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 964.046433] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Completed reading data from the image iterator. {{(pid=61728) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 964.046763] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234d3f3-005b-2d61-2f04-3ee093b98279/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 964.048080] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a1aee9-1870-4ed3-9d9b-14fd63f4a284 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.059162] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234d3f3-005b-2d61-2f04-3ee093b98279/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 964.059406] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234d3f3-005b-2d61-2f04-3ee093b98279/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 964.062438] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-b9794308-2e0a-4d64-9546-2d9822832ec5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.064120] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464567, 'name': Rename_Task, 'duration_secs': 0.380515} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.064403] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 964.064966] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad06968a-d993-4873-98c5-76c06f3db9bd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.070947] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 964.070947] env[61728]: value = "task-464569" [ 964.070947] env[61728]: _type = "Task" [ 964.070947] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.079113] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464569, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.117744] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464568, 'name': CloneVM_Task} progress is 93%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.364651] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "refresh_cache-db547fea-4d44-440e-ac3e-29ae9b2c547b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.364913] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired lock "refresh_cache-db547fea-4d44-440e-ac3e-29ae9b2c547b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.364946] env[61728]: DEBUG nova.network.neutron [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 964.465065] env[61728]: DEBUG oslo_vmware.rw_handles [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234d3f3-005b-2d61-2f04-3ee093b98279/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 964.465301] env[61728]: INFO nova.virt.vmwareapi.images [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Downloaded image file data 6bf918f6-e123-44bd-a072-fe70029d1786 [ 964.466153] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63293d51-a6c8-49f5-bdad-f7323f5135a3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.485290] env[61728]: DEBUG nova.compute.utils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 964.486952] env[61728]: DEBUG nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 964.487122] env[61728]: DEBUG nova.network.neutron [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 964.490769] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-62b170ad-07e9-46a8-aec6-186195144996 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.526325] env[61728]: INFO nova.virt.vmwareapi.images [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] The imported VM was unregistered [ 964.528894] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Caching image {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 964.528894] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Creating directory with path [datastore1] devstack-image-cache_base/6bf918f6-e123-44bd-a072-fe70029d1786 {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 964.529201] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-460a5bc9-0f5e-4efd-9cae-94d2e19c233d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.536985] env[61728]: DEBUG nova.policy [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8938cfb5439247d6b95adfe040babded', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9523aa656ff2459b94bb28d133fa4b05', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 964.555522] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Created directory with path [datastore1] devstack-image-cache_base/6bf918f6-e123-44bd-a072-fe70029d1786 {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 964.555730] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_85e605de-cdc3-4762-80b1-95d21bd5f31b/OSTACK_IMG_85e605de-cdc3-4762-80b1-95d21bd5f31b.vmdk to [datastore1] devstack-image-cache_base/6bf918f6-e123-44bd-a072-fe70029d1786/6bf918f6-e123-44bd-a072-fe70029d1786.vmdk. {{(pid=61728) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 964.556046] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-de7f2dbd-eb4a-454f-9fdb-7817e1383453 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.562796] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 964.562796] env[61728]: value = "task-464571" [ 964.562796] env[61728]: _type = "Task" [ 964.562796] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.570286] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464571, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.580386] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464569, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.618753] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464568, 'name': CloneVM_Task} progress is 93%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.771688] env[61728]: DEBUG nova.network.neutron [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Successfully created port: d3dfae3f-5d24-4189-8af1-a88d67ace2fa {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 964.911967] env[61728]: DEBUG nova.network.neutron [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 964.992765] env[61728]: DEBUG nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 965.081818] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464571, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.093539] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464569, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.130919] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464568, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.155794] env[61728]: DEBUG nova.network.neutron [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Updating instance_info_cache with network_info: [{"id": "94c4dd39-2357-472e-8680-6d65f1755e09", "address": "fa:16:3e:11:c0:df", "network": {"id": "74344aaa-7298-4ae0-ab9c-ebeef230b69d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1018009435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9523aa656ff2459b94bb28d133fa4b05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94c4dd39-23", "ovs_interfaceid": "94c4dd39-2357-472e-8680-6d65f1755e09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.390026] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e54d31-3256-4107-aee7-bfb5a376f941 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.397493] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c592bd8a-7198-4165-a34b-1c5ee719cb40 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.433252] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6a2ccf-bc9d-446d-98a6-90d9a3a33721 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.445025] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de65b6f-132a-4439-b041-3607dfda4e45 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.459723] env[61728]: DEBUG nova.compute.provider_tree [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.581986] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464571, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.589442] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464569, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.624797] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464568, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.663331] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Releasing lock "refresh_cache-db547fea-4d44-440e-ac3e-29ae9b2c547b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.663908] env[61728]: DEBUG nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Instance network_info: |[{"id": "94c4dd39-2357-472e-8680-6d65f1755e09", "address": "fa:16:3e:11:c0:df", "network": {"id": "74344aaa-7298-4ae0-ab9c-ebeef230b69d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1018009435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9523aa656ff2459b94bb28d133fa4b05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94c4dd39-23", "ovs_interfaceid": "94c4dd39-2357-472e-8680-6d65f1755e09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 965.665108] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:c0:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9bb629cd-6d0f-4bed-965c-bd04a2f3ec49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94c4dd39-2357-472e-8680-6d65f1755e09', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 965.679168] env[61728]: DEBUG oslo.service.loopingcall [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.680116] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 965.680518] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-55973ebc-7a84-4d64-83a8-02185ec8a27b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.718116] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.718116] env[61728]: value = "task-464572" [ 965.718116] env[61728]: _type = "Task" [ 965.718116] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.733804] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464572, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.820106] env[61728]: DEBUG nova.compute.manager [req-7820b167-5c65-4407-a16f-25012d5832ed req-27c9ba66-ddf8-409f-96bf-68acd9d3e90a service nova] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Received event network-changed-94c4dd39-2357-472e-8680-6d65f1755e09 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 965.820432] env[61728]: DEBUG nova.compute.manager [req-7820b167-5c65-4407-a16f-25012d5832ed req-27c9ba66-ddf8-409f-96bf-68acd9d3e90a service nova] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Refreshing instance network info cache due to event network-changed-94c4dd39-2357-472e-8680-6d65f1755e09. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 965.820624] env[61728]: DEBUG oslo_concurrency.lockutils [req-7820b167-5c65-4407-a16f-25012d5832ed req-27c9ba66-ddf8-409f-96bf-68acd9d3e90a service nova] Acquiring lock "refresh_cache-db547fea-4d44-440e-ac3e-29ae9b2c547b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.820797] env[61728]: DEBUG oslo_concurrency.lockutils [req-7820b167-5c65-4407-a16f-25012d5832ed req-27c9ba66-ddf8-409f-96bf-68acd9d3e90a service nova] Acquired lock "refresh_cache-db547fea-4d44-440e-ac3e-29ae9b2c547b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.820941] env[61728]: DEBUG nova.network.neutron [req-7820b167-5c65-4407-a16f-25012d5832ed req-27c9ba66-ddf8-409f-96bf-68acd9d3e90a service nova] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Refreshing network info cache for port 94c4dd39-2357-472e-8680-6d65f1755e09 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 965.963377] env[61728]: DEBUG nova.scheduler.client.report [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.005206] env[61728]: DEBUG nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 966.037422] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.037913] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.038162] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.038414] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.038600] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.038769] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.038996] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.039190] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.039366] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.039569] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.039786] env[61728]: DEBUG nova.virt.hardware [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.040740] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb1a52b8-47b1-4d46-a08f-2fce506aa0cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.051796] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a075c44b-7a2b-4322-b5e5-96e3e899ac3c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.080386] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464571, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.087726] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464569, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.123758] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464568, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.232460] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464572, 'name': CreateVM_Task, 'duration_secs': 0.481529} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.232645] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 966.233581] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.233904] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.234228] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.234537] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dafb6d15-8818-4d4a-b260-1a330e53fa7c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.242087] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 966.242087] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f4f7d6-060a-dbe9-a74b-2c5a798a1dd7" [ 966.242087] env[61728]: _type = "Task" [ 966.242087] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.253466] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f4f7d6-060a-dbe9-a74b-2c5a798a1dd7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.376754] env[61728]: DEBUG nova.network.neutron [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Successfully updated port: d3dfae3f-5d24-4189-8af1-a88d67ace2fa {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 966.468731] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.469435] env[61728]: DEBUG nova.compute.manager [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 966.472251] env[61728]: DEBUG oslo_concurrency.lockutils [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.340s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.472544] env[61728]: DEBUG nova.objects.instance [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lazy-loading 'resources' on Instance uuid 0e77ef23-3d1a-4f00-83f9-5b617334f980 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.573016] env[61728]: DEBUG nova.network.neutron [req-7820b167-5c65-4407-a16f-25012d5832ed req-27c9ba66-ddf8-409f-96bf-68acd9d3e90a service nova] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Updated VIF entry in instance network info cache for port 94c4dd39-2357-472e-8680-6d65f1755e09. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 966.573377] env[61728]: DEBUG nova.network.neutron [req-7820b167-5c65-4407-a16f-25012d5832ed req-27c9ba66-ddf8-409f-96bf-68acd9d3e90a service nova] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Updating instance_info_cache with network_info: [{"id": "94c4dd39-2357-472e-8680-6d65f1755e09", "address": "fa:16:3e:11:c0:df", "network": {"id": "74344aaa-7298-4ae0-ab9c-ebeef230b69d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1018009435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9523aa656ff2459b94bb28d133fa4b05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94c4dd39-23", "ovs_interfaceid": "94c4dd39-2357-472e-8680-6d65f1755e09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.582851] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464571, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.590320] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464569, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.628705] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464568, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.754301] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f4f7d6-060a-dbe9-a74b-2c5a798a1dd7, 'name': SearchDatastore_Task, 'duration_secs': 0.070783} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.754673] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.754962] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 966.755233] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.755389] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.755599] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.755912] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-062f22a1-ae82-42ff-b139-38d953fed699 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.773734] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 966.773958] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 966.775043] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec9a6d30-2884-4d74-b40f-3cba3f92a5bb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.782323] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 966.782323] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c7cb7f-576d-4a02-8565-0add0f645b98" [ 966.782323] env[61728]: _type = "Task" [ 966.782323] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.792544] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c7cb7f-576d-4a02-8565-0add0f645b98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.879762] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "refresh_cache-0ff847f9-97d0-40d9-a458-038fe36b5c62" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.879950] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired lock "refresh_cache-0ff847f9-97d0-40d9-a458-038fe36b5c62" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.880129] env[61728]: DEBUG nova.network.neutron [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 966.977080] env[61728]: DEBUG nova.compute.utils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 966.978317] env[61728]: DEBUG nova.compute.manager [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 966.978504] env[61728]: DEBUG nova.network.neutron [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 967.026215] env[61728]: DEBUG nova.policy [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb983aba45954b299f28bf8fc9389aad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f9e67e5c64a4cefbaafbeb8b0605011', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 967.082962] env[61728]: DEBUG oslo_concurrency.lockutils [req-7820b167-5c65-4407-a16f-25012d5832ed req-27c9ba66-ddf8-409f-96bf-68acd9d3e90a service nova] Releasing lock "refresh_cache-db547fea-4d44-440e-ac3e-29ae9b2c547b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.083477] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464571, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.092084] env[61728]: DEBUG oslo_vmware.api [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464569, 'name': PowerOnVM_Task, 'duration_secs': 2.776718} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.092375] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 967.092591] env[61728]: INFO nova.compute.manager [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Took 9.87 seconds to spawn the instance on the hypervisor. [ 967.092796] env[61728]: DEBUG nova.compute.manager [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.093635] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d626cf-497a-4648-965f-0739062644f7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.128502] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464568, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.301785] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c7cb7f-576d-4a02-8565-0add0f645b98, 'name': SearchDatastore_Task, 'duration_secs': 0.084478} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.301836] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0353299a-d595-4d39-ab91-e92f221f2f2c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.310338] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 967.310338] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52774b57-fb08-9196-cf5c-358498ff7090" [ 967.310338] env[61728]: _type = "Task" [ 967.310338] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.311367] env[61728]: DEBUG nova.network.neutron [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Successfully created port: 6fe3bf1e-7d6a-443d-ad35-b36d4000f769 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 967.326300] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52774b57-fb08-9196-cf5c-358498ff7090, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.341833] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e47c8ca-f61f-4f27-9bf6-16c4dc8199d0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.350462] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a2b607-4f43-4c4d-9a39-d514e393ea34 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.387462] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-128f7429-37c1-4f7b-bfcf-e305ba890570 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.398118] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabd2b06-ecc5-4630-91d2-a8136b8151ef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.414355] env[61728]: DEBUG nova.compute.provider_tree [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.431203] env[61728]: DEBUG nova.network.neutron [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 967.484353] env[61728]: DEBUG nova.compute.manager [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 967.585328] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464571, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.615761] env[61728]: INFO nova.compute.manager [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Took 43.36 seconds to build instance. [ 967.626179] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464568, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.647915] env[61728]: DEBUG nova.network.neutron [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Updating instance_info_cache with network_info: [{"id": "d3dfae3f-5d24-4189-8af1-a88d67ace2fa", "address": "fa:16:3e:d0:b0:00", "network": {"id": "74344aaa-7298-4ae0-ab9c-ebeef230b69d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1018009435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9523aa656ff2459b94bb28d133fa4b05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3dfae3f-5d", "ovs_interfaceid": "d3dfae3f-5d24-4189-8af1-a88d67ace2fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.821847] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52774b57-fb08-9196-cf5c-358498ff7090, 'name': SearchDatastore_Task, 'duration_secs': 0.086876} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.822148] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.822417] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] db547fea-4d44-440e-ac3e-29ae9b2c547b/db547fea-4d44-440e-ac3e-29ae9b2c547b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 967.822682] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f5b02c9a-6f23-4636-8605-38bf27aebd79 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.829337] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 967.829337] env[61728]: value = "task-464573" [ 967.829337] env[61728]: _type = "Task" [ 967.829337] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.837516] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464573, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.848307] env[61728]: DEBUG nova.compute.manager [req-b35e346c-70d2-4f69-a131-61856d21f0d8 req-d3413e31-46fc-4a85-bc56-a3ed8099a9d3 service nova] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Received event network-vif-plugged-d3dfae3f-5d24-4189-8af1-a88d67ace2fa {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 967.848307] env[61728]: DEBUG oslo_concurrency.lockutils [req-b35e346c-70d2-4f69-a131-61856d21f0d8 req-d3413e31-46fc-4a85-bc56-a3ed8099a9d3 service nova] Acquiring lock "0ff847f9-97d0-40d9-a458-038fe36b5c62-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.848307] env[61728]: DEBUG oslo_concurrency.lockutils [req-b35e346c-70d2-4f69-a131-61856d21f0d8 req-d3413e31-46fc-4a85-bc56-a3ed8099a9d3 service nova] Lock "0ff847f9-97d0-40d9-a458-038fe36b5c62-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.848453] env[61728]: DEBUG oslo_concurrency.lockutils [req-b35e346c-70d2-4f69-a131-61856d21f0d8 req-d3413e31-46fc-4a85-bc56-a3ed8099a9d3 service nova] Lock "0ff847f9-97d0-40d9-a458-038fe36b5c62-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.848625] env[61728]: DEBUG nova.compute.manager [req-b35e346c-70d2-4f69-a131-61856d21f0d8 req-d3413e31-46fc-4a85-bc56-a3ed8099a9d3 service nova] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] No waiting events found dispatching network-vif-plugged-d3dfae3f-5d24-4189-8af1-a88d67ace2fa {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 967.848798] env[61728]: WARNING nova.compute.manager [req-b35e346c-70d2-4f69-a131-61856d21f0d8 req-d3413e31-46fc-4a85-bc56-a3ed8099a9d3 service nova] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Received unexpected event network-vif-plugged-d3dfae3f-5d24-4189-8af1-a88d67ace2fa for instance with vm_state building and task_state spawning. [ 967.848966] env[61728]: DEBUG nova.compute.manager [req-b35e346c-70d2-4f69-a131-61856d21f0d8 req-d3413e31-46fc-4a85-bc56-a3ed8099a9d3 service nova] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Received event network-changed-d3dfae3f-5d24-4189-8af1-a88d67ace2fa {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 967.849144] env[61728]: DEBUG nova.compute.manager [req-b35e346c-70d2-4f69-a131-61856d21f0d8 req-d3413e31-46fc-4a85-bc56-a3ed8099a9d3 service nova] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Refreshing instance network info cache due to event network-changed-d3dfae3f-5d24-4189-8af1-a88d67ace2fa. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 967.849316] env[61728]: DEBUG oslo_concurrency.lockutils [req-b35e346c-70d2-4f69-a131-61856d21f0d8 req-d3413e31-46fc-4a85-bc56-a3ed8099a9d3 service nova] Acquiring lock "refresh_cache-0ff847f9-97d0-40d9-a458-038fe36b5c62" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.918271] env[61728]: DEBUG nova.scheduler.client.report [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.080261] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464571, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.394218} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.080602] env[61728]: INFO nova.virt.vmwareapi.ds_util [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_85e605de-cdc3-4762-80b1-95d21bd5f31b/OSTACK_IMG_85e605de-cdc3-4762-80b1-95d21bd5f31b.vmdk to [datastore1] devstack-image-cache_base/6bf918f6-e123-44bd-a072-fe70029d1786/6bf918f6-e123-44bd-a072-fe70029d1786.vmdk. [ 968.080742] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Cleaning up location [datastore1] OSTACK_IMG_85e605de-cdc3-4762-80b1-95d21bd5f31b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 968.080915] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_85e605de-cdc3-4762-80b1-95d21bd5f31b {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.081217] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b68f7d70-33a4-4392-a877-8ca971d50463 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.087406] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 968.087406] env[61728]: value = "task-464574" [ 968.087406] env[61728]: _type = "Task" [ 968.087406] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.095643] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464574, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.117226] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88cb1f8d-d5f8-4d41-9da4-b7c1240af9a7 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.869s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.125928] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464568, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.154026] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Releasing lock "refresh_cache-0ff847f9-97d0-40d9-a458-038fe36b5c62" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.154026] env[61728]: DEBUG nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Instance network_info: |[{"id": "d3dfae3f-5d24-4189-8af1-a88d67ace2fa", "address": "fa:16:3e:d0:b0:00", "network": {"id": "74344aaa-7298-4ae0-ab9c-ebeef230b69d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1018009435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9523aa656ff2459b94bb28d133fa4b05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3dfae3f-5d", "ovs_interfaceid": "d3dfae3f-5d24-4189-8af1-a88d67ace2fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 968.154026] env[61728]: DEBUG oslo_concurrency.lockutils [req-b35e346c-70d2-4f69-a131-61856d21f0d8 req-d3413e31-46fc-4a85-bc56-a3ed8099a9d3 service nova] Acquired lock "refresh_cache-0ff847f9-97d0-40d9-a458-038fe36b5c62" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.154026] env[61728]: DEBUG nova.network.neutron [req-b35e346c-70d2-4f69-a131-61856d21f0d8 req-d3413e31-46fc-4a85-bc56-a3ed8099a9d3 service nova] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Refreshing network info cache for port d3dfae3f-5d24-4189-8af1-a88d67ace2fa {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 968.154026] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:b0:00', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9bb629cd-6d0f-4bed-965c-bd04a2f3ec49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd3dfae3f-5d24-4189-8af1-a88d67ace2fa', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 968.165432] env[61728]: DEBUG oslo.service.loopingcall [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.166648] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 968.168548] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-491f177e-9b35-427d-b593-dac37b970e13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.188645] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 968.188645] env[61728]: value = "task-464575" [ 968.188645] env[61728]: _type = "Task" [ 968.188645] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.196624] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464575, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.339493] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464573, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.424322] env[61728]: DEBUG oslo_concurrency.lockutils [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.952s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.426759] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 23.393s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.454506] env[61728]: INFO nova.scheduler.client.report [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Deleted allocations for instance 0e77ef23-3d1a-4f00-83f9-5b617334f980 [ 968.493503] env[61728]: DEBUG nova.compute.manager [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 968.528382] env[61728]: DEBUG nova.virt.hardware [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 968.528730] env[61728]: DEBUG nova.virt.hardware [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 968.529012] env[61728]: DEBUG nova.virt.hardware [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 968.529227] env[61728]: DEBUG nova.virt.hardware [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 968.529383] env[61728]: DEBUG nova.virt.hardware [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 968.529538] env[61728]: DEBUG nova.virt.hardware [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 968.529764] env[61728]: DEBUG nova.virt.hardware [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 968.529934] env[61728]: DEBUG nova.virt.hardware [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 968.530130] env[61728]: DEBUG nova.virt.hardware [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 968.530309] env[61728]: DEBUG nova.virt.hardware [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 968.530492] env[61728]: DEBUG nova.virt.hardware [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 968.531975] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2baa38-6f38-4c5e-8fff-be9cfb341af2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.542705] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b7b8c9-a719-49f8-a4a7-a8eb27e38b60 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.600090] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464574, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.111501} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.600399] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.600624] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6bf918f6-e123-44bd-a072-fe70029d1786/6bf918f6-e123-44bd-a072-fe70029d1786.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.600892] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6bf918f6-e123-44bd-a072-fe70029d1786/6bf918f6-e123-44bd-a072-fe70029d1786.vmdk to [datastore1] 3bfefcfc-db97-4a9d-86cb-9fb1d8158863/3bfefcfc-db97-4a9d-86cb-9fb1d8158863.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 968.601197] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e61d96c9-b56a-4181-988f-04c1d23c480b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.610938] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 968.610938] env[61728]: value = "task-464576" [ 968.610938] env[61728]: _type = "Task" [ 968.610938] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.621152] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464576, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.630603] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464568, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.702703] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464575, 'name': CreateVM_Task, 'duration_secs': 0.382237} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.702909] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 968.703640] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.703813] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.704243] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 968.704525] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04f29d05-e732-46b8-bb48-444ee3fef7eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.712019] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 968.712019] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526828ac-e912-c02d-0d87-da04eb4cdb1b" [ 968.712019] env[61728]: _type = "Task" [ 968.712019] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.723469] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526828ac-e912-c02d-0d87-da04eb4cdb1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.838765] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464573, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.916591} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.839085] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] db547fea-4d44-440e-ac3e-29ae9b2c547b/db547fea-4d44-440e-ac3e-29ae9b2c547b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 968.839328] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 968.839616] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f0baf18-a5fd-491b-bd27-e88d727ef726 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.848708] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 968.848708] env[61728]: value = "task-464577" [ 968.848708] env[61728]: _type = "Task" [ 968.848708] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.857662] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464577, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.963311] env[61728]: DEBUG oslo_concurrency.lockutils [None req-47d0d4a6-94fb-4e59-81af-1291c6ef5282 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "0e77ef23-3d1a-4f00-83f9-5b617334f980" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.696s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.036795] env[61728]: DEBUG nova.network.neutron [req-b35e346c-70d2-4f69-a131-61856d21f0d8 req-d3413e31-46fc-4a85-bc56-a3ed8099a9d3 service nova] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Updated VIF entry in instance network info cache for port d3dfae3f-5d24-4189-8af1-a88d67ace2fa. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 969.037293] env[61728]: DEBUG nova.network.neutron [req-b35e346c-70d2-4f69-a131-61856d21f0d8 req-d3413e31-46fc-4a85-bc56-a3ed8099a9d3 service nova] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Updating instance_info_cache with network_info: [{"id": "d3dfae3f-5d24-4189-8af1-a88d67ace2fa", "address": "fa:16:3e:d0:b0:00", "network": {"id": "74344aaa-7298-4ae0-ab9c-ebeef230b69d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1018009435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9523aa656ff2459b94bb28d133fa4b05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3dfae3f-5d", "ovs_interfaceid": "d3dfae3f-5d24-4189-8af1-a88d67ace2fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.083704] env[61728]: DEBUG nova.compute.manager [req-af4c501d-c7ee-4904-a0c8-86a0be092e9b req-51b328da-b762-4ed3-8fc8-90d010665e14 service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Received event network-changed-67077c45-d87f-4d2d-a48b-d23a96d35f8a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 969.084041] env[61728]: DEBUG nova.compute.manager [req-af4c501d-c7ee-4904-a0c8-86a0be092e9b req-51b328da-b762-4ed3-8fc8-90d010665e14 service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Refreshing instance network info cache due to event network-changed-67077c45-d87f-4d2d-a48b-d23a96d35f8a. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 969.084193] env[61728]: DEBUG oslo_concurrency.lockutils [req-af4c501d-c7ee-4904-a0c8-86a0be092e9b req-51b328da-b762-4ed3-8fc8-90d010665e14 service nova] Acquiring lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.084342] env[61728]: DEBUG oslo_concurrency.lockutils [req-af4c501d-c7ee-4904-a0c8-86a0be092e9b req-51b328da-b762-4ed3-8fc8-90d010665e14 service nova] Acquired lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.084503] env[61728]: DEBUG nova.network.neutron [req-af4c501d-c7ee-4904-a0c8-86a0be092e9b req-51b328da-b762-4ed3-8fc8-90d010665e14 service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Refreshing network info cache for port 67077c45-d87f-4d2d-a48b-d23a96d35f8a {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 969.122871] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464576, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.132526] env[61728]: DEBUG nova.network.neutron [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Successfully updated port: 6fe3bf1e-7d6a-443d-ad35-b36d4000f769 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 969.137924] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464568, 'name': CloneVM_Task, 'duration_secs': 5.397721} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.139262] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Created linked-clone VM from snapshot [ 969.140071] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-844ac39a-b18b-47c4-a320-a45677587329 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.150725] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Uploading image 4b79277a-3323-457b-9c23-6de1a6d93646 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 969.168741] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 969.168741] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ee077f0d-bd58-4252-a1de-31d5f2c8a326 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.175466] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 969.175466] env[61728]: value = "task-464578" [ 969.175466] env[61728]: _type = "Task" [ 969.175466] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.186680] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464578, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.225720] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526828ac-e912-c02d-0d87-da04eb4cdb1b, 'name': SearchDatastore_Task, 'duration_secs': 0.022207} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.226299] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.226687] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 969.227050] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.227336] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.227626] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 969.228046] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cf6df119-ed97-4935-ab1e-153530635508 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.245584] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 969.246169] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 969.247079] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95118335-5df8-482b-9f64-c65ce46d5670 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.255011] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 969.255011] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dae497-52c8-45a1-2b4e-f40a93ff838f" [ 969.255011] env[61728]: _type = "Task" [ 969.255011] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.265959] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dae497-52c8-45a1-2b4e-f40a93ff838f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.362544] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464577, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113388} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.362882] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 969.363757] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb5b2c23-fa01-4612-ba57-752a0445976b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.388331] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] db547fea-4d44-440e-ac3e-29ae9b2c547b/db547fea-4d44-440e-ac3e-29ae9b2c547b.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 969.388701] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9abffa46-e2fe-4129-9ca6-56484f1b830f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.413150] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 969.413150] env[61728]: value = "task-464579" [ 969.413150] env[61728]: _type = "Task" [ 969.413150] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.424270] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464579, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.451423] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Applying migration context for instance 3415687e-9b8d-42f8-9b18-93951b8ec7da as it has an incoming, in-progress migration 92379aa6-c648-4aae-8141-f8deadfff31e. Migration status is confirming {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 969.454050] env[61728]: INFO nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating resource usage from migration 92379aa6-c648-4aae-8141-f8deadfff31e [ 969.454497] env[61728]: INFO nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating resource usage from migration 258316b5-7ba1-41d5-9080-956cb4a7720c [ 969.480683] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 40ef29df-12d6-4274-846a-4a2c32752b8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.482110] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 06e765e1-85d9-4f6c-8ab5-81735f08cb66 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 969.482110] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance f27f22aa-f05d-4981-a389-311731c51f93 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.482110] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 5fa2e9f4-6d68-46c2-a549-95d216bab886 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.482110] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance b1fb2e3f-234a-44c7-983e-d4441b4e3cef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.482110] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 2c166dd3-985c-4637-92ab-939b46a7c90e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.482110] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.482110] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 8c5b2380-bd14-451e-9612-2767c2a6adb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.482110] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance aebc4c7c-09d6-484f-8ad6-2416b3a41828 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 969.482110] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 3bfefcfc-db97-4a9d-86cb-9fb1d8158863 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.482450] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Migration 92379aa6-c648-4aae-8141-f8deadfff31e is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 969.482450] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 3415687e-9b8d-42f8-9b18-93951b8ec7da actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.482450] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance a70d1c37-e792-4168-b7e1-9418b8cb7818 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.482535] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 0af88a53-bbe7-443e-a7dc-734e483bb730 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.482582] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.482695] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance db547fea-4d44-440e-ac3e-29ae9b2c547b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.482832] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 0ff847f9-97d0-40d9-a458-038fe36b5c62 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.482953] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 7e8cde9c-890c-4f50-9ccb-46aa71220395 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.543798] env[61728]: DEBUG oslo_concurrency.lockutils [req-b35e346c-70d2-4f69-a131-61856d21f0d8 req-d3413e31-46fc-4a85-bc56-a3ed8099a9d3 service nova] Releasing lock "refresh_cache-0ff847f9-97d0-40d9-a458-038fe36b5c62" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.623982] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464576, 'name': CopyVirtualDisk_Task} progress is 32%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.632067] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Acquiring lock "refresh_cache-7e8cde9c-890c-4f50-9ccb-46aa71220395" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.632221] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Acquired lock "refresh_cache-7e8cde9c-890c-4f50-9ccb-46aa71220395" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.632372] env[61728]: DEBUG nova.network.neutron [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 969.690017] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464578, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.770111] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52dae497-52c8-45a1-2b4e-f40a93ff838f, 'name': SearchDatastore_Task, 'duration_secs': 0.072347} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.771095] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e92be0b-8fb4-4458-a1ac-c086aae9a626 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.778922] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 969.778922] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527c89c5-d2e4-b2ff-56a2-de91f6e80023" [ 969.778922] env[61728]: _type = "Task" [ 969.778922] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.789647] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527c89c5-d2e4-b2ff-56a2-de91f6e80023, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.845437] env[61728]: DEBUG nova.network.neutron [req-af4c501d-c7ee-4904-a0c8-86a0be092e9b req-51b328da-b762-4ed3-8fc8-90d010665e14 service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Updated VIF entry in instance network info cache for port 67077c45-d87f-4d2d-a48b-d23a96d35f8a. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 969.845916] env[61728]: DEBUG nova.network.neutron [req-af4c501d-c7ee-4904-a0c8-86a0be092e9b req-51b328da-b762-4ed3-8fc8-90d010665e14 service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Updating instance_info_cache with network_info: [{"id": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "address": "fa:16:3e:d1:31:88", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67077c45-d8", "ovs_interfaceid": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.925763] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464579, 'name': ReconfigVM_Task, 'duration_secs': 0.329549} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.925763] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Reconfigured VM instance instance-0000005d to attach disk [datastore1] db547fea-4d44-440e-ac3e-29ae9b2c547b/db547fea-4d44-440e-ac3e-29ae9b2c547b.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.926432] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-827fa90d-eb7c-4721-b7c0-3f2fa12fb6c4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.936545] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 969.936545] env[61728]: value = "task-464580" [ 969.936545] env[61728]: _type = "Task" [ 969.936545] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.951043] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464580, 'name': Rename_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.975401] env[61728]: DEBUG nova.compute.manager [req-c8248b75-18c0-4a44-a059-b670cc913480 req-0a5f79b3-72b2-4371-8469-569e376e878a service nova] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Received event network-vif-plugged-6fe3bf1e-7d6a-443d-ad35-b36d4000f769 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 969.975575] env[61728]: DEBUG oslo_concurrency.lockutils [req-c8248b75-18c0-4a44-a059-b670cc913480 req-0a5f79b3-72b2-4371-8469-569e376e878a service nova] Acquiring lock "7e8cde9c-890c-4f50-9ccb-46aa71220395-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.975890] env[61728]: DEBUG oslo_concurrency.lockutils [req-c8248b75-18c0-4a44-a059-b670cc913480 req-0a5f79b3-72b2-4371-8469-569e376e878a service nova] Lock "7e8cde9c-890c-4f50-9ccb-46aa71220395-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.976098] env[61728]: DEBUG oslo_concurrency.lockutils [req-c8248b75-18c0-4a44-a059-b670cc913480 req-0a5f79b3-72b2-4371-8469-569e376e878a service nova] Lock "7e8cde9c-890c-4f50-9ccb-46aa71220395-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.976333] env[61728]: DEBUG nova.compute.manager [req-c8248b75-18c0-4a44-a059-b670cc913480 req-0a5f79b3-72b2-4371-8469-569e376e878a service nova] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] No waiting events found dispatching network-vif-plugged-6fe3bf1e-7d6a-443d-ad35-b36d4000f769 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 969.976549] env[61728]: WARNING nova.compute.manager [req-c8248b75-18c0-4a44-a059-b670cc913480 req-0a5f79b3-72b2-4371-8469-569e376e878a service nova] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Received unexpected event network-vif-plugged-6fe3bf1e-7d6a-443d-ad35-b36d4000f769 for instance with vm_state building and task_state spawning. [ 969.976745] env[61728]: DEBUG nova.compute.manager [req-c8248b75-18c0-4a44-a059-b670cc913480 req-0a5f79b3-72b2-4371-8469-569e376e878a service nova] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Received event network-changed-6fe3bf1e-7d6a-443d-ad35-b36d4000f769 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 969.977031] env[61728]: DEBUG nova.compute.manager [req-c8248b75-18c0-4a44-a059-b670cc913480 req-0a5f79b3-72b2-4371-8469-569e376e878a service nova] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Refreshing instance network info cache due to event network-changed-6fe3bf1e-7d6a-443d-ad35-b36d4000f769. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 969.977210] env[61728]: DEBUG oslo_concurrency.lockutils [req-c8248b75-18c0-4a44-a059-b670cc913480 req-0a5f79b3-72b2-4371-8469-569e376e878a service nova] Acquiring lock "refresh_cache-7e8cde9c-890c-4f50-9ccb-46aa71220395" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.986750] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 705c8838-f605-46fa-b036-ec212f673f66 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 969.987072] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Migration 258316b5-7ba1-41d5-9080-956cb4a7720c is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 969.987234] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance f9f07464-9f43-43fd-8895-2673861747fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.127986] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464576, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.167863] env[61728]: DEBUG nova.network.neutron [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 970.189178] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464578, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.293343] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527c89c5-d2e4-b2ff-56a2-de91f6e80023, 'name': SearchDatastore_Task, 'duration_secs': 0.081251} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.296113] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.296423] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 0ff847f9-97d0-40d9-a458-038fe36b5c62/0ff847f9-97d0-40d9-a458-038fe36b5c62.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 970.296716] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-606cbebc-7b00-4c4a-baac-298212cbc01e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.306152] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 970.306152] env[61728]: value = "task-464581" [ 970.306152] env[61728]: _type = "Task" [ 970.306152] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.309939] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "2c166dd3-985c-4637-92ab-939b46a7c90e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.310320] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "2c166dd3-985c-4637-92ab-939b46a7c90e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.310430] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "2c166dd3-985c-4637-92ab-939b46a7c90e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.310616] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "2c166dd3-985c-4637-92ab-939b46a7c90e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.310786] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "2c166dd3-985c-4637-92ab-939b46a7c90e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.313213] env[61728]: INFO nova.compute.manager [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Terminating instance [ 970.315489] env[61728]: DEBUG nova.compute.manager [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 970.315703] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 970.316633] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e3bc459-da06-4a3d-8c7f-3bb843584d09 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.326995] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464581, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.329403] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 970.329686] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0713ecbd-7016-4124-97b9-5f1100142055 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.335844] env[61728]: DEBUG oslo_vmware.api [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 970.335844] env[61728]: value = "task-464582" [ 970.335844] env[61728]: _type = "Task" [ 970.335844] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.340281] env[61728]: DEBUG nova.network.neutron [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Updating instance_info_cache with network_info: [{"id": "6fe3bf1e-7d6a-443d-ad35-b36d4000f769", "address": "fa:16:3e:99:0e:89", "network": {"id": "36f6fe38-500d-49b2-885c-4c282d31ca68", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-600888795-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f9e67e5c64a4cefbaafbeb8b0605011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fe3bf1e-7d", "ovs_interfaceid": "6fe3bf1e-7d6a-443d-ad35-b36d4000f769", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.349711] env[61728]: DEBUG oslo_concurrency.lockutils [req-af4c501d-c7ee-4904-a0c8-86a0be092e9b req-51b328da-b762-4ed3-8fc8-90d010665e14 service nova] Releasing lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.351141] env[61728]: DEBUG oslo_vmware.api [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464582, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.450219] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.450549] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.450804] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.451115] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.451339] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.453976] env[61728]: INFO nova.compute.manager [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Terminating instance [ 970.456032] env[61728]: DEBUG nova.compute.manager [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 970.456251] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 970.456680] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464580, 'name': Rename_Task, 'duration_secs': 0.154764} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.457785] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959799db-37ca-42b0-8f61-8739d5ab61db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.461764] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 970.461764] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a117f065-21cc-4e21-997a-21e64cd9293b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.470079] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 970.471437] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58de3a33-d5c0-469c-9219-a1ca15323735 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.473221] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 970.473221] env[61728]: value = "task-464583" [ 970.473221] env[61728]: _type = "Task" [ 970.473221] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.480069] env[61728]: DEBUG oslo_vmware.api [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 970.480069] env[61728]: value = "task-464584" [ 970.480069] env[61728]: _type = "Task" [ 970.480069] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.487265] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464583, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.491596] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 970.497578] env[61728]: DEBUG oslo_vmware.api [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464584, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.625105] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464576, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.690210] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464578, 'name': Destroy_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.819819] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464581, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.844068] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Releasing lock "refresh_cache-7e8cde9c-890c-4f50-9ccb-46aa71220395" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.844068] env[61728]: DEBUG nova.compute.manager [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Instance network_info: |[{"id": "6fe3bf1e-7d6a-443d-ad35-b36d4000f769", "address": "fa:16:3e:99:0e:89", "network": {"id": "36f6fe38-500d-49b2-885c-4c282d31ca68", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-600888795-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f9e67e5c64a4cefbaafbeb8b0605011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fe3bf1e-7d", "ovs_interfaceid": "6fe3bf1e-7d6a-443d-ad35-b36d4000f769", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 970.844407] env[61728]: DEBUG oslo_concurrency.lockutils [req-c8248b75-18c0-4a44-a059-b670cc913480 req-0a5f79b3-72b2-4371-8469-569e376e878a service nova] Acquired lock "refresh_cache-7e8cde9c-890c-4f50-9ccb-46aa71220395" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.844553] env[61728]: DEBUG nova.network.neutron [req-c8248b75-18c0-4a44-a059-b670cc913480 req-0a5f79b3-72b2-4371-8469-569e376e878a service nova] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Refreshing network info cache for port 6fe3bf1e-7d6a-443d-ad35-b36d4000f769 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 970.846268] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:0e:89', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a31c4b8-5b72-4f32-aab3-c4e963e684dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6fe3bf1e-7d6a-443d-ad35-b36d4000f769', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 970.858928] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Creating folder: Project (4f9e67e5c64a4cefbaafbeb8b0605011). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 970.867634] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-688b78f7-c4cc-4289-8ab0-53ada9b07a69 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.869821] env[61728]: DEBUG oslo_vmware.api [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464582, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.880284] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Created folder: Project (4f9e67e5c64a4cefbaafbeb8b0605011) in parent group-v121913. [ 970.880500] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Creating folder: Instances. Parent ref: group-v122171. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 970.880757] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c0ceb69-8232-4fd0-9e74-f147b77daeb8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.892070] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Created folder: Instances in parent group-v122171. [ 970.892365] env[61728]: DEBUG oslo.service.loopingcall [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.892579] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 970.892809] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-310c0e1c-91a2-4d84-bb24-903fbbb0a4e3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.917578] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 970.917578] env[61728]: value = "task-464587" [ 970.917578] env[61728]: _type = "Task" [ 970.917578] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.925596] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464587, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.983942] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464583, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.997266] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 172129ff-136e-4855-8659-3f1ac88dcbd2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 970.997515] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Total usable vcpus: 48, total allocated vcpus: 17 {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 970.997665] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3840MB phys_disk=200GB used_disk=17GB total_vcpus=48 used_vcpus=17 pci_stats=[] {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 971.000172] env[61728]: DEBUG oslo_vmware.api [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464584, 'name': PowerOffVM_Task, 'duration_secs': 0.427066} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.000677] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 971.000869] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 971.001398] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-079b9762-a82e-4e17-8cce-1c8a8df53bf5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.071053] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 971.071314] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 971.071505] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Deleting the datastore file [datastore1] ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.071832] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c5bc4ea-bef3-41db-ab51-a0d3ae3bd3e3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.086175] env[61728]: DEBUG oslo_vmware.api [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 971.086175] env[61728]: value = "task-464589" [ 971.086175] env[61728]: _type = "Task" [ 971.086175] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.095189] env[61728]: DEBUG oslo_vmware.api [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464589, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.116452] env[61728]: DEBUG nova.network.neutron [req-c8248b75-18c0-4a44-a059-b670cc913480 req-0a5f79b3-72b2-4371-8469-569e376e878a service nova] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Updated VIF entry in instance network info cache for port 6fe3bf1e-7d6a-443d-ad35-b36d4000f769. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 971.116530] env[61728]: DEBUG nova.network.neutron [req-c8248b75-18c0-4a44-a059-b670cc913480 req-0a5f79b3-72b2-4371-8469-569e376e878a service nova] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Updating instance_info_cache with network_info: [{"id": "6fe3bf1e-7d6a-443d-ad35-b36d4000f769", "address": "fa:16:3e:99:0e:89", "network": {"id": "36f6fe38-500d-49b2-885c-4c282d31ca68", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-600888795-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f9e67e5c64a4cefbaafbeb8b0605011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fe3bf1e-7d", "ovs_interfaceid": "6fe3bf1e-7d6a-443d-ad35-b36d4000f769", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.129355] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464576, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.193855] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464578, 'name': Destroy_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.319806] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464581, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.348055] env[61728]: DEBUG oslo_vmware.api [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464582, 'name': PowerOffVM_Task, 'duration_secs': 0.947274} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.348395] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 971.348637] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 971.348879] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1163685a-8586-48f5-92e1-666d2d9caaea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.351958] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c008c6-7764-41a4-a35d-6681e784a4b1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.359862] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4200c63-3851-4677-b8ec-d367d70cfcc5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.396947] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0684d39-7685-4e87-b3be-efdb9b74c872 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.405235] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98c47a9-f240-4b21-8c5d-4bb16415e90f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.420139] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.433743] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464587, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.446883] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 971.447181] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 971.447391] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Deleting the datastore file [datastore1] 2c166dd3-985c-4637-92ab-939b46a7c90e {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.447748] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd98caea-424f-41c0-97e5-2d298e8a61c8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.456692] env[61728]: DEBUG oslo_vmware.api [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for the task: (returnval){ [ 971.456692] env[61728]: value = "task-464591" [ 971.456692] env[61728]: _type = "Task" [ 971.456692] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.466926] env[61728]: DEBUG oslo_vmware.api [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464591, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.486791] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464583, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.597174] env[61728]: DEBUG oslo_vmware.api [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464589, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.624944] env[61728]: DEBUG oslo_concurrency.lockutils [req-c8248b75-18c0-4a44-a059-b670cc913480 req-0a5f79b3-72b2-4371-8469-569e376e878a service nova] Releasing lock "refresh_cache-7e8cde9c-890c-4f50-9ccb-46aa71220395" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.634262] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464576, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.710848} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.635573] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6bf918f6-e123-44bd-a072-fe70029d1786/6bf918f6-e123-44bd-a072-fe70029d1786.vmdk to [datastore1] 3bfefcfc-db97-4a9d-86cb-9fb1d8158863/3bfefcfc-db97-4a9d-86cb-9fb1d8158863.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 971.637075] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177877e8-cdbe-44b1-b375-204e047636ef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.686612] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] 3bfefcfc-db97-4a9d-86cb-9fb1d8158863/3bfefcfc-db97-4a9d-86cb-9fb1d8158863.vmdk or device None with type streamOptimized {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.687320] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1932e604-5426-4252-b3e8-5a7a4e70591c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.722340] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 971.722340] env[61728]: value = "task-464592" [ 971.722340] env[61728]: _type = "Task" [ 971.722340] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.726180] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464578, 'name': Destroy_Task, 'duration_secs': 2.225327} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.730160] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Destroyed the VM [ 971.730534] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 971.730900] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-751029a2-5916-4a46-9a4a-90cc791d817c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.739615] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464592, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.741891] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 971.741891] env[61728]: value = "task-464593" [ 971.741891] env[61728]: _type = "Task" [ 971.741891] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.751246] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464593, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.818992] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464581, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.475761} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.819407] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 0ff847f9-97d0-40d9-a458-038fe36b5c62/0ff847f9-97d0-40d9-a458-038fe36b5c62.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 971.819665] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 971.819941] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-17c4c70f-72e4-42ad-9e14-bb48ef728f64 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.826442] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 971.826442] env[61728]: value = "task-464594" [ 971.826442] env[61728]: _type = "Task" [ 971.826442] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.836730] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464594, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.930148] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.933929] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464587, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.966823] env[61728]: DEBUG oslo_vmware.api [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464591, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.983878] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464583, 'name': PowerOnVM_Task, 'duration_secs': 1.343849} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.984305] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 971.984549] env[61728]: INFO nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Took 8.44 seconds to spawn the instance on the hypervisor. [ 971.984751] env[61728]: DEBUG nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 971.985528] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d1d354-7f94-4147-91f6-e71525adad1c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.097399] env[61728]: DEBUG oslo_vmware.api [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464589, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.726086} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.097695] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.097880] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 972.098100] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 972.098286] env[61728]: INFO nova.compute.manager [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Took 1.64 seconds to destroy the instance on the hypervisor. [ 972.098530] env[61728]: DEBUG oslo.service.loopingcall [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.098727] env[61728]: DEBUG nova.compute.manager [-] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.098840] env[61728]: DEBUG nova.network.neutron [-] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 972.239438] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464592, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.251655] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464593, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.337929] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464594, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.19448} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.338245] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 972.339361] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5db5268-73d6-4b3b-ba6b-692bc5982463 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.361954] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 0ff847f9-97d0-40d9-a458-038fe36b5c62/0ff847f9-97d0-40d9-a458-038fe36b5c62.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 972.363720] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1a1ab0c-119d-4eaf-af30-03d729a4cda2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.380846] env[61728]: DEBUG nova.compute.manager [req-8646f400-6baa-4e1f-b977-458cee8c2dfc req-9538500f-bf9c-4e28-bf3a-ef779a9f632d service nova] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Received event network-vif-deleted-05dc7f10-50f6-4a11-95b8-a6bbd35821b5 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 972.381058] env[61728]: INFO nova.compute.manager [req-8646f400-6baa-4e1f-b977-458cee8c2dfc req-9538500f-bf9c-4e28-bf3a-ef779a9f632d service nova] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Neutron deleted interface 05dc7f10-50f6-4a11-95b8-a6bbd35821b5; detaching it from the instance and deleting it from the info cache [ 972.381242] env[61728]: DEBUG nova.network.neutron [req-8646f400-6baa-4e1f-b977-458cee8c2dfc req-9538500f-bf9c-4e28-bf3a-ef779a9f632d service nova] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.388860] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 972.388860] env[61728]: value = "task-464595" [ 972.388860] env[61728]: _type = "Task" [ 972.388860] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.398491] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464595, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.430116] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464587, 'name': CreateVM_Task, 'duration_secs': 1.099531} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.430302] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 972.430971] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.431165] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.431507] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 972.431758] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbc8dfdc-49cb-472e-969a-c8f3e5f500ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.435356] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 972.435582] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.009s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.436913] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.206s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.437174] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.439229] env[61728]: DEBUG oslo_concurrency.lockutils [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.534s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.439419] env[61728]: DEBUG oslo_concurrency.lockutils [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.441378] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 17.115s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.442992] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Waiting for the task: (returnval){ [ 972.442992] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520cbf77-5b9c-1e51-0dbc-df39fab60195" [ 972.442992] env[61728]: _type = "Task" [ 972.442992] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.452145] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520cbf77-5b9c-1e51-0dbc-df39fab60195, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.465916] env[61728]: INFO nova.scheduler.client.report [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Deleted allocations for instance 06e765e1-85d9-4f6c-8ab5-81735f08cb66 [ 972.467307] env[61728]: INFO nova.scheduler.client.report [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleted allocations for instance aebc4c7c-09d6-484f-8ad6-2416b3a41828 [ 972.471025] env[61728]: DEBUG oslo_vmware.api [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Task: {'id': task-464591, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.669558} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.475102] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.475283] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 972.475465] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 972.475642] env[61728]: INFO nova.compute.manager [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Took 2.16 seconds to destroy the instance on the hypervisor. [ 972.475879] env[61728]: DEBUG oslo.service.loopingcall [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.476527] env[61728]: DEBUG nova.compute.manager [-] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.476630] env[61728]: DEBUG nova.network.neutron [-] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 972.501362] env[61728]: INFO nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Took 42.35 seconds to build instance. [ 972.737981] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464592, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.752477] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464593, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.879185] env[61728]: DEBUG nova.network.neutron [-] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.885037] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-295bee90-8374-4368-a5e7-3ba0241d5f3a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.897922] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013f0170-ad53-4505-97cc-a52962138332 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.912271] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464595, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.937494] env[61728]: DEBUG nova.compute.manager [req-8646f400-6baa-4e1f-b977-458cee8c2dfc req-9538500f-bf9c-4e28-bf3a-ef779a9f632d service nova] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Detach interface failed, port_id=05dc7f10-50f6-4a11-95b8-a6bbd35821b5, reason: Instance ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 972.956223] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520cbf77-5b9c-1e51-0dbc-df39fab60195, 'name': SearchDatastore_Task, 'duration_secs': 0.024822} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.956610] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.957211] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 972.957211] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.957211] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.957394] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 972.957685] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a130edf7-7b09-4c48-aa8d-6aa12a08eff3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.969163] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 972.969374] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 972.970420] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c80da05b-cf8e-4633-bbca-00a27717efda {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.976469] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Waiting for the task: (returnval){ [ 972.976469] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5213f88a-8164-7c8c-cf50-fa75b5746c5d" [ 972.976469] env[61728]: _type = "Task" [ 972.976469] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.986493] env[61728]: DEBUG oslo_concurrency.lockutils [None req-95f40e78-fd29-40f5-a46a-81237fd543d4 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "06e765e1-85d9-4f6c-8ab5-81735f08cb66" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.510s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.987704] env[61728]: DEBUG oslo_concurrency.lockutils [None req-03b6daa7-2371-470d-ac7a-54f5d1d2c926 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "aebc4c7c-09d6-484f-8ad6-2416b3a41828" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.687s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.994192] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5213f88a-8164-7c8c-cf50-fa75b5746c5d, 'name': SearchDatastore_Task, 'duration_secs': 0.012007} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.995059] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ebc36df-6133-4f92-b868-b4b6a0b64cb5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.003307] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "db547fea-4d44-440e-ac3e-29ae9b2c547b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.863s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.003637] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Waiting for the task: (returnval){ [ 973.003637] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5218c8da-fa8b-fd0c-310c-9fa3fec7f3f8" [ 973.003637] env[61728]: _type = "Task" [ 973.003637] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.018569] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5218c8da-fa8b-fd0c-310c-9fa3fec7f3f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.195396] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d041467-0db6-4279-89c0-a3881777a780 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.198535] env[61728]: DEBUG nova.network.neutron [-] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.205369] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e706b2-b902-4b5b-9b66-5d43b44131d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.242109] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bbf688b-43f5-40c9-aed9-ebfa1b428712 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.253941] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464592, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.257356] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e93c90-675b-4de9-a49c-17687966ff47 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.265408] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464593, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.277367] env[61728]: DEBUG nova.compute.provider_tree [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.381841] env[61728]: INFO nova.compute.manager [-] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Took 1.28 seconds to deallocate network for instance. [ 973.400741] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464595, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.518025] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5218c8da-fa8b-fd0c-310c-9fa3fec7f3f8, 'name': SearchDatastore_Task, 'duration_secs': 0.04856} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.518025] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.518025] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 7e8cde9c-890c-4f50-9ccb-46aa71220395/7e8cde9c-890c-4f50-9ccb-46aa71220395.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 973.518025] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dce69806-80cd-4db2-92f3-4feb9fda0e5f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.525869] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Waiting for the task: (returnval){ [ 973.525869] env[61728]: value = "task-464596" [ 973.525869] env[61728]: _type = "Task" [ 973.525869] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.539589] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464596, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.703906] env[61728]: INFO nova.compute.manager [-] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Took 1.22 seconds to deallocate network for instance. [ 973.752590] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464592, 'name': ReconfigVM_Task, 'duration_secs': 1.650452} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.756018] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Reconfigured VM instance instance-0000002e to attach disk [datastore1] 3bfefcfc-db97-4a9d-86cb-9fb1d8158863/3bfefcfc-db97-4a9d-86cb-9fb1d8158863.vmdk or device None with type streamOptimized {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.756018] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-43956bd1-bf29-4aa8-968e-bef4ca9ae5ee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.760142] env[61728]: DEBUG oslo_vmware.api [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464593, 'name': RemoveSnapshot_Task, 'duration_secs': 1.746098} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.760952] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 973.769016] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 973.769016] env[61728]: value = "task-464597" [ 973.769016] env[61728]: _type = "Task" [ 973.769016] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.777493] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464597, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.781852] env[61728]: DEBUG nova.scheduler.client.report [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 973.890118] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.900367] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464595, 'name': ReconfigVM_Task, 'duration_secs': 1.194048} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.900727] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 0ff847f9-97d0-40d9-a458-038fe36b5c62/0ff847f9-97d0-40d9-a458-038fe36b5c62.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.901471] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0148bc89-5436-4237-9259-33787cba1e2e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.911474] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 973.911474] env[61728]: value = "task-464598" [ 973.911474] env[61728]: _type = "Task" [ 973.911474] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.924028] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464598, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.037553] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464596, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.211255] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.265811] env[61728]: WARNING nova.compute.manager [None req-0ba3f241-c63c-4799-b1e6-71d4a6aad0b7 tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Image not found during snapshot: nova.exception.ImageNotFound: Image 4b79277a-3323-457b-9c23-6de1a6d93646 could not be found. [ 974.280517] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464597, 'name': Rename_Task, 'duration_secs': 0.16372} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.280813] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 974.282038] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-977bbf63-b789-4204-bd22-3254af09cf2e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.294040] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 974.294040] env[61728]: value = "task-464599" [ 974.294040] env[61728]: _type = "Task" [ 974.294040] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.303203] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464599, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.405129] env[61728]: DEBUG nova.compute.manager [req-7832c339-c850-4dc8-93a2-6ec38f86daf0 req-42997970-8713-418d-8a43-cee12b65e093 service nova] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Received event network-vif-deleted-bcd321c5-6557-4c39-a5e4-025a4b20d7db {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 974.422518] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464598, 'name': Rename_Task, 'duration_secs': 0.180967} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.422960] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 974.423285] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32988b06-167c-43f4-bc7d-0e2ed99d0e74 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.430772] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 974.430772] env[61728]: value = "task-464600" [ 974.430772] env[61728]: _type = "Task" [ 974.430772] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.448737] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464600, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.539637] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464596, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.635161} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.539998] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 7e8cde9c-890c-4f50-9ccb-46aa71220395/7e8cde9c-890c-4f50-9ccb-46aa71220395.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 974.540305] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 974.540789] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b6c272e-fd66-4b08-963e-61a78cd175c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.550668] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Waiting for the task: (returnval){ [ 974.550668] env[61728]: value = "task-464601" [ 974.550668] env[61728]: _type = "Task" [ 974.550668] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.560582] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464601, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.604069] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "b4241d6f-1d70-4f61-a322-caff7c4e6f64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.604456] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "b4241d6f-1d70-4f61-a322-caff7c4e6f64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.800283] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.357s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.802869] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.312s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.804912] env[61728]: INFO nova.compute.claims [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 974.814421] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464599, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.944328] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464600, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.060642] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464601, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086198} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.060977] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 975.061798] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c193d44b-9947-41b0-a335-3496d33ba4b2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.084055] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 7e8cde9c-890c-4f50-9ccb-46aa71220395/7e8cde9c-890c-4f50-9ccb-46aa71220395.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.084949] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "0af88a53-bbe7-443e-a7dc-734e483bb730" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.084949] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "0af88a53-bbe7-443e-a7dc-734e483bb730" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.084949] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "0af88a53-bbe7-443e-a7dc-734e483bb730-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.085137] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "0af88a53-bbe7-443e-a7dc-734e483bb730-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.085315] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "0af88a53-bbe7-443e-a7dc-734e483bb730-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.086910] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19c415b6-6b4a-4891-875b-3c7aa6c1a6bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.101831] env[61728]: INFO nova.compute.manager [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Terminating instance [ 975.106994] env[61728]: DEBUG nova.compute.manager [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 975.112291] env[61728]: DEBUG nova.compute.manager [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 975.112291] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 975.112291] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Waiting for the task: (returnval){ [ 975.112291] env[61728]: value = "task-464602" [ 975.112291] env[61728]: _type = "Task" [ 975.112291] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.112990] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad7c6e96-9ad3-4652-a4d5-6067142d6320 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.123611] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 975.126963] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5035c20-f94f-4744-bb4a-4a295adcee72 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.128826] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464602, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.135926] env[61728]: DEBUG oslo_vmware.api [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 975.135926] env[61728]: value = "task-464603" [ 975.135926] env[61728]: _type = "Task" [ 975.135926] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.146511] env[61728]: DEBUG oslo_vmware.api [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464603, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.308443] env[61728]: DEBUG oslo_vmware.api [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464599, 'name': PowerOnVM_Task, 'duration_secs': 0.554422} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.319786] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 975.389470] env[61728]: INFO nova.scheduler.client.report [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleted allocation for migration 92379aa6-c648-4aae-8141-f8deadfff31e [ 975.432068] env[61728]: DEBUG nova.compute.manager [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.433150] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-168a0dc8-fae8-41a7-9fa9-80c01cfe5c97 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.453425] env[61728]: DEBUG oslo_vmware.api [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464600, 'name': PowerOnVM_Task, 'duration_secs': 0.617539} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.454234] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 975.454699] env[61728]: INFO nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Took 9.45 seconds to spawn the instance on the hypervisor. [ 975.455053] env[61728]: DEBUG nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.456388] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89937f2f-ab4f-46d9-a7bc-5a10b5f6a4ee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.630963] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464602, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.634656] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.647390] env[61728]: DEBUG oslo_vmware.api [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464603, 'name': PowerOffVM_Task, 'duration_secs': 0.362237} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.647712] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 975.647860] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 975.648136] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-151b19e6-e985-4da1-94b3-51e5a996287d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.721160] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 975.721396] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 975.721586] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleting the datastore file [datastore1] 0af88a53-bbe7-443e-a7dc-734e483bb730 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 975.721886] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0ac36572-5e11-4622-bd89-5ab4858ab332 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.729521] env[61728]: DEBUG oslo_vmware.api [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for the task: (returnval){ [ 975.729521] env[61728]: value = "task-464605" [ 975.729521] env[61728]: _type = "Task" [ 975.729521] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.737961] env[61728]: DEBUG oslo_vmware.api [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464605, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.896696] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ccb9568e-6067-450f-8dcc-d0d115a038f0 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "3415687e-9b8d-42f8-9b18-93951b8ec7da" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 23.943s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.965047] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d3d9d9ce-3fe6-43a3-b7b5-5b4d19a3c56c tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 52.884s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.977676] env[61728]: INFO nova.compute.manager [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Took 45.80 seconds to build instance. [ 976.134700] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464602, 'name': ReconfigVM_Task, 'duration_secs': 0.756578} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.134789] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 7e8cde9c-890c-4f50-9ccb-46aa71220395/7e8cde9c-890c-4f50-9ccb-46aa71220395.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.135439] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-24192720-a96b-4689-a4fc-975105ea383f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.145697] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "267880ff-0fac-4763-82be-ef8587ee7628" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.146137] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "267880ff-0fac-4763-82be-ef8587ee7628" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.149878] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Waiting for the task: (returnval){ [ 976.149878] env[61728]: value = "task-464606" [ 976.149878] env[61728]: _type = "Task" [ 976.149878] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.163954] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464606, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.208658] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73656f93-1e1a-45ad-a1c2-d6bf9a8afce1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.218083] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e8b2249-a250-455f-806c-78de65b02a1b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.254474] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186b277f-4c9d-4136-850d-f930cbaae3d2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.262790] env[61728]: DEBUG oslo_vmware.api [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Task: {'id': task-464605, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.349528} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.264958] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 976.265178] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 976.265362] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 976.265545] env[61728]: INFO nova.compute.manager [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Took 1.15 seconds to destroy the instance on the hypervisor. [ 976.265793] env[61728]: DEBUG oslo.service.loopingcall [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.266066] env[61728]: DEBUG nova.compute.manager [-] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 976.266164] env[61728]: DEBUG nova.network.neutron [-] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 976.268758] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c206c36-f9b2-4027-8b1c-90e1ee5eafa3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.284138] env[61728]: DEBUG nova.compute.provider_tree [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.484741] env[61728]: DEBUG oslo_concurrency.lockutils [None req-60f8909b-b798-45ec-ae22-9f0e207590d2 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "0ff847f9-97d0-40d9-a458-038fe36b5c62" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.313s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.655028] env[61728]: DEBUG nova.compute.manager [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 976.665049] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464606, 'name': Rename_Task, 'duration_secs': 0.267647} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.665722] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 976.666111] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-044eb692-d262-4c60-8db8-6434f00a7308 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.674901] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Waiting for the task: (returnval){ [ 976.674901] env[61728]: value = "task-464607" [ 976.674901] env[61728]: _type = "Task" [ 976.674901] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.685064] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464607, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.752654] env[61728]: DEBUG nova.compute.manager [req-586aa4c0-0b61-42ec-8729-d8d80cb0b3be req-efaa5070-9ea2-4007-98e9-c66c34eae933 service nova] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Received event network-vif-deleted-814bcd52-6c0c-4efd-ae00-1cfb3375cd63 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 976.756014] env[61728]: INFO nova.compute.manager [req-586aa4c0-0b61-42ec-8729-d8d80cb0b3be req-efaa5070-9ea2-4007-98e9-c66c34eae933 service nova] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Neutron deleted interface 814bcd52-6c0c-4efd-ae00-1cfb3375cd63; detaching it from the instance and deleting it from the info cache [ 976.756014] env[61728]: DEBUG nova.network.neutron [req-586aa4c0-0b61-42ec-8729-d8d80cb0b3be req-efaa5070-9ea2-4007-98e9-c66c34eae933 service nova] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.786857] env[61728]: DEBUG nova.scheduler.client.report [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.814446] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "3415687e-9b8d-42f8-9b18-93951b8ec7da" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.814664] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "3415687e-9b8d-42f8-9b18-93951b8ec7da" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.814907] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "3415687e-9b8d-42f8-9b18-93951b8ec7da-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.815108] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "3415687e-9b8d-42f8-9b18-93951b8ec7da-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.815356] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "3415687e-9b8d-42f8-9b18-93951b8ec7da-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.818690] env[61728]: INFO nova.compute.manager [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Terminating instance [ 976.824996] env[61728]: DEBUG nova.compute.manager [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 976.825225] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 976.826598] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bacebbff-58d0-4acd-8bda-47ecb5277558 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.839395] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 976.839690] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a3b3d39-44d0-46a9-961b-8296b68ac407 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.848631] env[61728]: DEBUG oslo_vmware.api [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 976.848631] env[61728]: value = "task-464608" [ 976.848631] env[61728]: _type = "Task" [ 976.848631] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.859801] env[61728]: DEBUG oslo_vmware.api [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464608, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.019372] env[61728]: DEBUG nova.network.neutron [-] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.180535] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.187563] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464607, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.258791] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d8daccd8-3368-409c-9701-dc27394bdaf4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.270561] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a6c99d-dc58-4733-99c7-e576a00bcd14 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.293075] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.293624] env[61728]: DEBUG nova.compute.manager [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 977.309022] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 18.826s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.311433] env[61728]: DEBUG nova.compute.manager [req-586aa4c0-0b61-42ec-8729-d8d80cb0b3be req-efaa5070-9ea2-4007-98e9-c66c34eae933 service nova] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Detach interface failed, port_id=814bcd52-6c0c-4efd-ae00-1cfb3375cd63, reason: Instance 0af88a53-bbe7-443e-a7dc-734e483bb730 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 977.325657] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "db547fea-4d44-440e-ac3e-29ae9b2c547b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.325981] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "db547fea-4d44-440e-ac3e-29ae9b2c547b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.326242] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "db547fea-4d44-440e-ac3e-29ae9b2c547b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.326532] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "db547fea-4d44-440e-ac3e-29ae9b2c547b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.326756] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "db547fea-4d44-440e-ac3e-29ae9b2c547b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.329811] env[61728]: INFO nova.compute.manager [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Terminating instance [ 977.332081] env[61728]: DEBUG nova.compute.manager [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 977.332352] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 977.333233] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b06c6b-397a-4f45-bc75-12751d8925cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.342176] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 977.342464] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e3f94bad-b774-44b4-a921-e1244fb59755 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.350467] env[61728]: DEBUG oslo_vmware.api [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 977.350467] env[61728]: value = "task-464609" [ 977.350467] env[61728]: _type = "Task" [ 977.350467] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.362524] env[61728]: DEBUG oslo_vmware.api [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464608, 'name': PowerOffVM_Task, 'duration_secs': 0.185139} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.366029] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 977.366263] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 977.366449] env[61728]: DEBUG oslo_vmware.api [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464609, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.366708] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-872d15ca-c4de-43d0-9925-fb95c1f2c0ee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.387349] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "378ef752-6d47-4861-baf1-064b2a180dcc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.387594] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "378ef752-6d47-4861-baf1-064b2a180dcc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.522478] env[61728]: INFO nova.compute.manager [-] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Took 1.26 seconds to deallocate network for instance. [ 977.530945] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "0ff847f9-97d0-40d9-a458-038fe36b5c62" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.531224] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "0ff847f9-97d0-40d9-a458-038fe36b5c62" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.531464] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "0ff847f9-97d0-40d9-a458-038fe36b5c62-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.531724] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "0ff847f9-97d0-40d9-a458-038fe36b5c62-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.531916] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "0ff847f9-97d0-40d9-a458-038fe36b5c62-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.533893] env[61728]: INFO nova.compute.manager [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Terminating instance [ 977.535994] env[61728]: DEBUG nova.compute.manager [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 977.536438] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 977.537118] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75d20cc7-183f-4ca9-b82d-5d11fd8471be {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.545821] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 977.547022] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3cda54f8-f325-43ce-995a-4a4cc7e365ec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.552890] env[61728]: DEBUG oslo_vmware.api [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 977.552890] env[61728]: value = "task-464611" [ 977.552890] env[61728]: _type = "Task" [ 977.552890] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.563081] env[61728]: DEBUG oslo_vmware.api [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464611, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.686763] env[61728]: DEBUG oslo_vmware.api [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464607, 'name': PowerOnVM_Task, 'duration_secs': 0.529293} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.687100] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 977.687323] env[61728]: INFO nova.compute.manager [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Took 9.19 seconds to spawn the instance on the hypervisor. [ 977.687512] env[61728]: DEBUG nova.compute.manager [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 977.688312] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e437aa-e2e3-4140-9a47-d49ce51cb77d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.814487] env[61728]: DEBUG nova.compute.utils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 977.818013] env[61728]: INFO nova.compute.claims [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.822556] env[61728]: DEBUG nova.compute.manager [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 977.822556] env[61728]: DEBUG nova.network.neutron [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 977.866409] env[61728]: DEBUG oslo_vmware.api [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464609, 'name': PowerOffVM_Task, 'duration_secs': 0.214161} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.867892] env[61728]: DEBUG nova.policy [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e51efcf23db4d5496edbc3b7104b4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32bb35e1dfce40e48be08bb568d3f2b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 977.869570] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 977.869749] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 977.870354] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-70213935-acfb-4f0e-97bd-5ec76d9bf4b6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.891292] env[61728]: DEBUG nova.compute.manager [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 978.029048] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.063604] env[61728]: DEBUG oslo_vmware.api [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464611, 'name': PowerOffVM_Task, 'duration_secs': 0.38178} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.064105] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 978.064311] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 978.064619] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-018dbcf9-2004-4eb9-b195-7cc9117c72a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.179913] env[61728]: DEBUG nova.network.neutron [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Successfully created port: ec0ae08d-62b0-4506-bc0b-65e9152dd5e8 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 978.207270] env[61728]: INFO nova.compute.manager [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Took 42.32 seconds to build instance. [ 978.324803] env[61728]: DEBUG nova.compute.manager [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 978.328028] env[61728]: INFO nova.compute.resource_tracker [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating resource usage from migration 258316b5-7ba1-41d5-9080-956cb4a7720c [ 978.350164] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f9b2b8-08fc-467b-a9ff-9449331dda52 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.360114] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b92b55b3-a88f-4360-b97a-72eac7ae8ed2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Suspending the VM {{(pid=61728) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 978.360114] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-496430b6-1929-460e-8cff-994550313eba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.370105] env[61728]: DEBUG oslo_vmware.api [None req-b92b55b3-a88f-4360-b97a-72eac7ae8ed2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 978.370105] env[61728]: value = "task-464614" [ 978.370105] env[61728]: _type = "Task" [ 978.370105] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.385663] env[61728]: DEBUG oslo_vmware.api [None req-b92b55b3-a88f-4360-b97a-72eac7ae8ed2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464614, 'name': SuspendVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.410430] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.643774] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49bd632f-3caa-4514-ad7c-ef5dc8c8662a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.652646] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-492cddd4-9142-413b-8d6a-66365f77d061 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.683767] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e562a79d-e787-4408-bb32-3f443877ba12 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.692671] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023aa1b2-a354-43db-adaf-27cb06bdb42f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.707453] env[61728]: DEBUG nova.compute.provider_tree [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.708964] env[61728]: DEBUG oslo_concurrency.lockutils [None req-14cadb39-c72f-4f82-b099-852d266bf937 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Lock "7e8cde9c-890c-4f50-9ccb-46aa71220395" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.829s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.881673] env[61728]: DEBUG oslo_vmware.api [None req-b92b55b3-a88f-4360-b97a-72eac7ae8ed2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464614, 'name': SuspendVM_Task} progress is 62%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.210658] env[61728]: DEBUG nova.scheduler.client.report [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.214486] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Acquiring lock "7e8cde9c-890c-4f50-9ccb-46aa71220395" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.214730] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Lock "7e8cde9c-890c-4f50-9ccb-46aa71220395" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.214936] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Acquiring lock "7e8cde9c-890c-4f50-9ccb-46aa71220395-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.215138] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Lock "7e8cde9c-890c-4f50-9ccb-46aa71220395-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.215309] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Lock "7e8cde9c-890c-4f50-9ccb-46aa71220395-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.219456] env[61728]: INFO nova.compute.manager [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Terminating instance [ 979.221422] env[61728]: DEBUG nova.compute.manager [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 979.221851] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 979.222773] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302bb1a9-c395-42dd-b71a-3259d0a79b29 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.234048] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 979.234315] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25981727-76a2-4941-8dd0-26fa5bffc2af {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.241046] env[61728]: DEBUG oslo_vmware.api [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Waiting for the task: (returnval){ [ 979.241046] env[61728]: value = "task-464615" [ 979.241046] env[61728]: _type = "Task" [ 979.241046] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.249415] env[61728]: DEBUG oslo_vmware.api [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464615, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.337033] env[61728]: DEBUG nova.compute.manager [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 979.363822] env[61728]: DEBUG nova.virt.hardware [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.363912] env[61728]: DEBUG nova.virt.hardware [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.364026] env[61728]: DEBUG nova.virt.hardware [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.364234] env[61728]: DEBUG nova.virt.hardware [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.364396] env[61728]: DEBUG nova.virt.hardware [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.364551] env[61728]: DEBUG nova.virt.hardware [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.364769] env[61728]: DEBUG nova.virt.hardware [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.365151] env[61728]: DEBUG nova.virt.hardware [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.365208] env[61728]: DEBUG nova.virt.hardware [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.365382] env[61728]: DEBUG nova.virt.hardware [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.365653] env[61728]: DEBUG nova.virt.hardware [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.366933] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3470a332-3345-4646-8c75-a41fe9ff2768 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.378156] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e7353b-b317-4d08-8474-50257e6ea1bd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.385368] env[61728]: DEBUG oslo_vmware.api [None req-b92b55b3-a88f-4360-b97a-72eac7ae8ed2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464614, 'name': SuspendVM_Task, 'duration_secs': 0.649191} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.386105] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b92b55b3-a88f-4360-b97a-72eac7ae8ed2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Suspended the VM {{(pid=61728) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 979.386375] env[61728]: DEBUG nova.compute.manager [None req-b92b55b3-a88f-4360-b97a-72eac7ae8ed2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 979.387124] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4b1e3c-0fec-4b28-90ba-03121af967c4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.500381] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 979.500624] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 979.500799] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Deleting the datastore file [datastore1] db547fea-4d44-440e-ac3e-29ae9b2c547b {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.502880] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba6847e7-d6d4-40b2-8ef3-1cd3431a3920 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.504809] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 979.505012] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 979.505201] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleting the datastore file [datastore1] 3415687e-9b8d-42f8-9b18-93951b8ec7da {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.505482] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 979.505672] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 979.505840] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Deleting the datastore file [datastore1] 0ff847f9-97d0-40d9-a458-038fe36b5c62 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.506135] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62b786f9-15d8-45d3-99d0-e6db5c742515 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.508203] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a88e431-2402-47c5-adb2-d23d8e2962d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.517853] env[61728]: DEBUG oslo_vmware.api [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 979.517853] env[61728]: value = "task-464617" [ 979.517853] env[61728]: _type = "Task" [ 979.517853] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.519356] env[61728]: DEBUG oslo_vmware.api [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 979.519356] env[61728]: value = "task-464616" [ 979.519356] env[61728]: _type = "Task" [ 979.519356] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.519616] env[61728]: DEBUG oslo_vmware.api [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for the task: (returnval){ [ 979.519616] env[61728]: value = "task-464618" [ 979.519616] env[61728]: _type = "Task" [ 979.519616] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.536113] env[61728]: DEBUG oslo_vmware.api [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464617, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.536381] env[61728]: DEBUG oslo_vmware.api [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464616, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.539127] env[61728]: DEBUG oslo_vmware.api [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464618, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.584206] env[61728]: DEBUG nova.compute.manager [req-74a2a458-61f3-477e-86fe-c15341f0410e req-093f972f-8bcd-46c3-b356-367ed2df0205 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Received event network-vif-plugged-ec0ae08d-62b0-4506-bc0b-65e9152dd5e8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 979.584575] env[61728]: DEBUG oslo_concurrency.lockutils [req-74a2a458-61f3-477e-86fe-c15341f0410e req-093f972f-8bcd-46c3-b356-367ed2df0205 service nova] Acquiring lock "705c8838-f605-46fa-b036-ec212f673f66-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.585093] env[61728]: DEBUG oslo_concurrency.lockutils [req-74a2a458-61f3-477e-86fe-c15341f0410e req-093f972f-8bcd-46c3-b356-367ed2df0205 service nova] Lock "705c8838-f605-46fa-b036-ec212f673f66-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.585485] env[61728]: DEBUG oslo_concurrency.lockutils [req-74a2a458-61f3-477e-86fe-c15341f0410e req-093f972f-8bcd-46c3-b356-367ed2df0205 service nova] Lock "705c8838-f605-46fa-b036-ec212f673f66-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.585894] env[61728]: DEBUG nova.compute.manager [req-74a2a458-61f3-477e-86fe-c15341f0410e req-093f972f-8bcd-46c3-b356-367ed2df0205 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] No waiting events found dispatching network-vif-plugged-ec0ae08d-62b0-4506-bc0b-65e9152dd5e8 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 979.586644] env[61728]: WARNING nova.compute.manager [req-74a2a458-61f3-477e-86fe-c15341f0410e req-093f972f-8bcd-46c3-b356-367ed2df0205 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Received unexpected event network-vif-plugged-ec0ae08d-62b0-4506-bc0b-65e9152dd5e8 for instance with vm_state building and task_state spawning. [ 979.662397] env[61728]: DEBUG nova.network.neutron [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Successfully updated port: ec0ae08d-62b0-4506-bc0b-65e9152dd5e8 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 979.715808] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.407s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.716189] env[61728]: INFO nova.compute.manager [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Migrating [ 979.722857] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.981s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.724675] env[61728]: INFO nova.compute.claims [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 979.752100] env[61728]: DEBUG oslo_vmware.api [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464615, 'name': PowerOffVM_Task, 'duration_secs': 0.20682} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.752409] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 979.752626] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 979.752903] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e769d8b5-4d41-4f52-a556-63891ea8d103 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.815239] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 979.815494] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 979.815684] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Deleting the datastore file [datastore1] 7e8cde9c-890c-4f50-9ccb-46aa71220395 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.815959] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95d609ba-9475-4fe5-a806-11249903927b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.823522] env[61728]: DEBUG oslo_vmware.api [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Waiting for the task: (returnval){ [ 979.823522] env[61728]: value = "task-464620" [ 979.823522] env[61728]: _type = "Task" [ 979.823522] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.831993] env[61728]: DEBUG oslo_vmware.api [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464620, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.031924] env[61728]: DEBUG oslo_vmware.api [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464617, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176032} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.032527] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 980.032724] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 980.032964] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 980.033178] env[61728]: INFO nova.compute.manager [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Took 3.21 seconds to destroy the instance on the hypervisor. [ 980.033431] env[61728]: DEBUG oslo.service.loopingcall [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.033964] env[61728]: DEBUG nova.compute.manager [-] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 980.034101] env[61728]: DEBUG nova.network.neutron [-] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 980.041108] env[61728]: DEBUG oslo_vmware.api [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464618, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171482} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.041772] env[61728]: DEBUG oslo_vmware.api [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Task: {'id': task-464616, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166754} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.041880] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 980.042082] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 980.042272] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 980.042512] env[61728]: INFO nova.compute.manager [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Took 2.51 seconds to destroy the instance on the hypervisor. [ 980.042708] env[61728]: DEBUG oslo.service.loopingcall [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.042902] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 980.043092] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 980.043270] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 980.043435] env[61728]: INFO nova.compute.manager [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Took 2.71 seconds to destroy the instance on the hypervisor. [ 980.043639] env[61728]: DEBUG oslo.service.loopingcall [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.043825] env[61728]: DEBUG nova.compute.manager [-] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 980.043917] env[61728]: DEBUG nova.network.neutron [-] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 980.045345] env[61728]: DEBUG nova.compute.manager [-] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 980.045446] env[61728]: DEBUG nova.network.neutron [-] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 980.165694] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.165694] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.165694] env[61728]: DEBUG nova.network.neutron [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 980.237395] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.237395] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.237395] env[61728]: DEBUG nova.network.neutron [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 980.335071] env[61728]: DEBUG oslo_vmware.api [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Task: {'id': task-464620, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150843} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.336035] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 980.336416] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 980.336887] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 980.337248] env[61728]: INFO nova.compute.manager [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Took 1.12 seconds to destroy the instance on the hypervisor. [ 980.339058] env[61728]: DEBUG oslo.service.loopingcall [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.339058] env[61728]: DEBUG nova.compute.manager [-] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 980.339058] env[61728]: DEBUG nova.network.neutron [-] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 980.697834] env[61728]: DEBUG nova.network.neutron [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 980.781986] env[61728]: DEBUG nova.compute.manager [req-924f1314-3aa1-48fc-b5bc-1aca26c9fc9b req-5d344e22-09f5-42d2-87a1-dc6125dd6330 service nova] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Received event network-vif-deleted-6fe3bf1e-7d6a-443d-ad35-b36d4000f769 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 980.782206] env[61728]: INFO nova.compute.manager [req-924f1314-3aa1-48fc-b5bc-1aca26c9fc9b req-5d344e22-09f5-42d2-87a1-dc6125dd6330 service nova] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Neutron deleted interface 6fe3bf1e-7d6a-443d-ad35-b36d4000f769; detaching it from the instance and deleting it from the info cache [ 980.782395] env[61728]: DEBUG nova.network.neutron [req-924f1314-3aa1-48fc-b5bc-1aca26c9fc9b req-5d344e22-09f5-42d2-87a1-dc6125dd6330 service nova] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.842023] env[61728]: DEBUG nova.network.neutron [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Updating instance_info_cache with network_info: [{"id": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "address": "fa:16:3e:57:58:2b", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec0ae08d-62", "ovs_interfaceid": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.053519] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9877f710-4b20-4ccf-9437-96536f4c7eda {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.059798] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694e5c35-6d30-4207-a9ea-156fd141ec48 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.091134] env[61728]: DEBUG nova.network.neutron [-] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.092720] env[61728]: DEBUG nova.network.neutron [-] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.096572] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2a05a8-816a-4d17-b6cb-de65c23f86af {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.103848] env[61728]: DEBUG nova.network.neutron [-] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.109174] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c163898e-e2d3-46f2-8e0b-6049f53a3ebc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.127039] env[61728]: DEBUG nova.compute.provider_tree [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 981.155750] env[61728]: DEBUG nova.network.neutron [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance_info_cache with network_info: [{"id": "83657468-5cac-4258-84f4-b105395d9cfa", "address": "fa:16:3e:55:1a:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83657468-5c", "ovs_interfaceid": "83657468-5cac-4258-84f4-b105395d9cfa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.176225] env[61728]: DEBUG nova.network.neutron [-] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.262195] env[61728]: INFO nova.compute.manager [None req-22412f64-92b3-420b-9dd5-cd73d36cdb4b tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Resuming [ 981.262875] env[61728]: DEBUG nova.objects.instance [None req-22412f64-92b3-420b-9dd5-cd73d36cdb4b tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lazy-loading 'flavor' on Instance uuid 3bfefcfc-db97-4a9d-86cb-9fb1d8158863 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.287526] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-041121c8-6ece-4b03-9817-790196505e26 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.298933] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2abec6b4-7eec-40b7-b9c3-34d55a126291 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.332181] env[61728]: DEBUG nova.compute.manager [req-924f1314-3aa1-48fc-b5bc-1aca26c9fc9b req-5d344e22-09f5-42d2-87a1-dc6125dd6330 service nova] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Detach interface failed, port_id=6fe3bf1e-7d6a-443d-ad35-b36d4000f769, reason: Instance 7e8cde9c-890c-4f50-9ccb-46aa71220395 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 981.345221] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.345605] env[61728]: DEBUG nova.compute.manager [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Instance network_info: |[{"id": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "address": "fa:16:3e:57:58:2b", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec0ae08d-62", "ovs_interfaceid": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 981.346533] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:58:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec0ae08d-62b0-4506-bc0b-65e9152dd5e8', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 981.355217] env[61728]: DEBUG oslo.service.loopingcall [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.355419] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 981.355648] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc3a4502-e47d-4ba3-b8cf-63032198c294 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.377353] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 981.377353] env[61728]: value = "task-464621" [ 981.377353] env[61728]: _type = "Task" [ 981.377353] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.385338] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464621, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.594910] env[61728]: INFO nova.compute.manager [-] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Took 1.55 seconds to deallocate network for instance. [ 981.600490] env[61728]: INFO nova.compute.manager [-] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Took 1.55 seconds to deallocate network for instance. [ 981.606351] env[61728]: INFO nova.compute.manager [-] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Took 1.57 seconds to deallocate network for instance. [ 981.620530] env[61728]: DEBUG nova.compute.manager [req-6c62755f-8706-4c67-b843-1cff6d86905a req-b703acb4-6989-4d5e-9b82-cb6d616145f2 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Received event network-changed-ec0ae08d-62b0-4506-bc0b-65e9152dd5e8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 981.620530] env[61728]: DEBUG nova.compute.manager [req-6c62755f-8706-4c67-b843-1cff6d86905a req-b703acb4-6989-4d5e-9b82-cb6d616145f2 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Refreshing instance network info cache due to event network-changed-ec0ae08d-62b0-4506-bc0b-65e9152dd5e8. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 981.620530] env[61728]: DEBUG oslo_concurrency.lockutils [req-6c62755f-8706-4c67-b843-1cff6d86905a req-b703acb4-6989-4d5e-9b82-cb6d616145f2 service nova] Acquiring lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.620643] env[61728]: DEBUG oslo_concurrency.lockutils [req-6c62755f-8706-4c67-b843-1cff6d86905a req-b703acb4-6989-4d5e-9b82-cb6d616145f2 service nova] Acquired lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.620800] env[61728]: DEBUG nova.network.neutron [req-6c62755f-8706-4c67-b843-1cff6d86905a req-b703acb4-6989-4d5e-9b82-cb6d616145f2 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Refreshing network info cache for port ec0ae08d-62b0-4506-bc0b-65e9152dd5e8 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 981.649391] env[61728]: ERROR nova.scheduler.client.report [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [req-d4d5fe71-940c-49fa-a443-a9915c179b4e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d4d5fe71-940c-49fa-a443-a9915c179b4e"}]} [ 981.658102] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.666745] env[61728]: DEBUG nova.scheduler.client.report [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 981.678593] env[61728]: INFO nova.compute.manager [-] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Took 1.34 seconds to deallocate network for instance. [ 981.683943] env[61728]: DEBUG nova.scheduler.client.report [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 981.684210] env[61728]: DEBUG nova.compute.provider_tree [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 981.698274] env[61728]: DEBUG nova.scheduler.client.report [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 981.716919] env[61728]: DEBUG nova.scheduler.client.report [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 981.891109] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464621, 'name': CreateVM_Task, 'duration_secs': 0.31268} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.893593] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 981.894536] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.894705] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.895036] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 981.895302] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45e80fd5-df13-45e8-801f-66ed2a7b5723 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.900531] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 981.900531] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c4e7b6-a4b7-d966-30b0-e175ba40ae76" [ 981.900531] env[61728]: _type = "Task" [ 981.900531] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.913300] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c4e7b6-a4b7-d966-30b0-e175ba40ae76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.936320] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d563ed11-004d-4a9a-afa3-017d856edb3d tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.936542] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d563ed11-004d-4a9a-afa3-017d856edb3d tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.936683] env[61728]: DEBUG nova.compute.manager [None req-d563ed11-004d-4a9a-afa3-017d856edb3d tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 981.939621] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741fcd96-3e83-4bb5-9cd0-cc46c4a082e1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.946572] env[61728]: DEBUG nova.compute.manager [None req-d563ed11-004d-4a9a-afa3-017d856edb3d tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61728) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 981.947152] env[61728]: DEBUG nova.objects.instance [None req-d563ed11-004d-4a9a-afa3-017d856edb3d tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lazy-loading 'flavor' on Instance uuid 8c5b2380-bd14-451e-9612-2767c2a6adb2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.998366] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981ce030-8579-409e-af92-f3743075b7a4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.006500] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf9b42e-2397-466f-8872-cfcbadfcffb9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.038657] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946b70be-ace7-45af-bac1-13de2f6333ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.046601] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58cdc9bd-df6d-4df0-8ad5-72ca659b99d3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.059704] env[61728]: DEBUG nova.compute.provider_tree [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 982.101385] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.107550] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.111499] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.188520] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.274176] env[61728]: DEBUG oslo_concurrency.lockutils [None req-22412f64-92b3-420b-9dd5-cd73d36cdb4b tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.274407] env[61728]: DEBUG oslo_concurrency.lockutils [None req-22412f64-92b3-420b-9dd5-cd73d36cdb4b tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquired lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.274635] env[61728]: DEBUG nova.network.neutron [None req-22412f64-92b3-420b-9dd5-cd73d36cdb4b tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 982.340205] env[61728]: DEBUG nova.network.neutron [req-6c62755f-8706-4c67-b843-1cff6d86905a req-b703acb4-6989-4d5e-9b82-cb6d616145f2 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Updated VIF entry in instance network info cache for port ec0ae08d-62b0-4506-bc0b-65e9152dd5e8. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 982.340581] env[61728]: DEBUG nova.network.neutron [req-6c62755f-8706-4c67-b843-1cff6d86905a req-b703acb4-6989-4d5e-9b82-cb6d616145f2 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Updating instance_info_cache with network_info: [{"id": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "address": "fa:16:3e:57:58:2b", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec0ae08d-62", "ovs_interfaceid": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.412018] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c4e7b6-a4b7-d966-30b0-e175ba40ae76, 'name': SearchDatastore_Task, 'duration_secs': 0.011516} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.412340] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.412587] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 982.412869] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.413044] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.413235] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 982.413547] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a98cfac1-964d-439c-989e-e033f0474d94 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.422531] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 982.422800] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 982.423535] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8370ef33-4022-4ed6-a70d-cd8d230c88a0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.428993] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 982.428993] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52154b17-36a5-960c-90b7-404727cf3abd" [ 982.428993] env[61728]: _type = "Task" [ 982.428993] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.436442] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52154b17-36a5-960c-90b7-404727cf3abd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.452086] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d563ed11-004d-4a9a-afa3-017d856edb3d tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 982.452318] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-667c6511-c84f-450a-b1b6-2322578d2e45 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.459272] env[61728]: DEBUG oslo_vmware.api [None req-d563ed11-004d-4a9a-afa3-017d856edb3d tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 982.459272] env[61728]: value = "task-464622" [ 982.459272] env[61728]: _type = "Task" [ 982.459272] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.469070] env[61728]: DEBUG oslo_vmware.api [None req-d563ed11-004d-4a9a-afa3-017d856edb3d tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464622, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.591575] env[61728]: DEBUG nova.scheduler.client.report [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 149 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 982.591961] env[61728]: DEBUG nova.compute.provider_tree [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 149 to 150 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 982.592175] env[61728]: DEBUG nova.compute.provider_tree [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 982.843175] env[61728]: DEBUG oslo_concurrency.lockutils [req-6c62755f-8706-4c67-b843-1cff6d86905a req-b703acb4-6989-4d5e-9b82-cb6d616145f2 service nova] Releasing lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.843518] env[61728]: DEBUG nova.compute.manager [req-6c62755f-8706-4c67-b843-1cff6d86905a req-b703acb4-6989-4d5e-9b82-cb6d616145f2 service nova] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Received event network-vif-deleted-d3dfae3f-5d24-4189-8af1-a88d67ace2fa {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 982.843639] env[61728]: DEBUG nova.compute.manager [req-6c62755f-8706-4c67-b843-1cff6d86905a req-b703acb4-6989-4d5e-9b82-cb6d616145f2 service nova] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Received event network-vif-deleted-94c4dd39-2357-472e-8680-6d65f1755e09 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 982.843896] env[61728]: DEBUG nova.compute.manager [req-6c62755f-8706-4c67-b843-1cff6d86905a req-b703acb4-6989-4d5e-9b82-cb6d616145f2 service nova] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Received event network-vif-deleted-fbf3382b-a343-4cc4-9233-ebf4ac3139a4 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 982.942943] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52154b17-36a5-960c-90b7-404727cf3abd, 'name': SearchDatastore_Task, 'duration_secs': 0.008063} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.943778] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9de476ae-aad5-4ed6-8122-998505852305 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.949177] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 982.949177] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52893782-7202-c7ca-0898-f315d00bd2e7" [ 982.949177] env[61728]: _type = "Task" [ 982.949177] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.957896] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52893782-7202-c7ca-0898-f315d00bd2e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.967891] env[61728]: DEBUG oslo_vmware.api [None req-d563ed11-004d-4a9a-afa3-017d856edb3d tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464622, 'name': PowerOffVM_Task, 'duration_secs': 0.209274} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.968194] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d563ed11-004d-4a9a-afa3-017d856edb3d tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 982.968380] env[61728]: DEBUG nova.compute.manager [None req-d563ed11-004d-4a9a-afa3-017d856edb3d tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 982.969109] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387c5924-0938-4051-9ac1-753d9336526d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.012377] env[61728]: DEBUG nova.network.neutron [None req-22412f64-92b3-420b-9dd5-cd73d36cdb4b tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Updating instance_info_cache with network_info: [{"id": "124d4780-15ec-4f17-a2c4-edf1517ed589", "address": "fa:16:3e:6d:47:93", "network": {"id": "6a1908cc-f6aa-4dd7-970e-c8a07e1b16ee", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2047209328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "01839050e56e4dbfa6e22471a3c60998", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap124d4780-15", "ovs_interfaceid": "124d4780-15ec-4f17-a2c4-edf1517ed589", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.097773] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.375s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.098535] env[61728]: DEBUG nova.compute.manager [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 983.102355] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.516s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.105035] env[61728]: INFO nova.compute.claims [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 983.174306] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ecbed2-0132-48e1-bb10-2129c37795fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.196555] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance 'f9f07464-9f43-43fd-8895-2673861747fa' progress to 0 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 983.461399] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52893782-7202-c7ca-0898-f315d00bd2e7, 'name': SearchDatastore_Task, 'duration_secs': 0.01047} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.461727] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.461972] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 705c8838-f605-46fa-b036-ec212f673f66/705c8838-f605-46fa-b036-ec212f673f66.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 983.462262] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d70da98-7248-42c1-aad4-d59a4c7529fa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.470776] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 983.470776] env[61728]: value = "task-464623" [ 983.470776] env[61728]: _type = "Task" [ 983.470776] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.481362] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464623, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.484999] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d563ed11-004d-4a9a-afa3-017d856edb3d tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.548s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.515393] env[61728]: DEBUG oslo_concurrency.lockutils [None req-22412f64-92b3-420b-9dd5-cd73d36cdb4b tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Releasing lock "refresh_cache-3bfefcfc-db97-4a9d-86cb-9fb1d8158863" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.516387] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72864812-1086-4257-8046-7f661a2fc141 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.528050] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-22412f64-92b3-420b-9dd5-cd73d36cdb4b tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Resuming the VM {{(pid=61728) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 983.528150] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0452b34c-0690-4e17-8932-903d51333439 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.538375] env[61728]: DEBUG oslo_vmware.api [None req-22412f64-92b3-420b-9dd5-cd73d36cdb4b tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 983.538375] env[61728]: value = "task-464624" [ 983.538375] env[61728]: _type = "Task" [ 983.538375] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.547156] env[61728]: DEBUG oslo_vmware.api [None req-22412f64-92b3-420b-9dd5-cd73d36cdb4b tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464624, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.610170] env[61728]: DEBUG nova.compute.utils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 983.611747] env[61728]: DEBUG nova.compute.manager [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 983.611939] env[61728]: DEBUG nova.network.neutron [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 983.670893] env[61728]: DEBUG nova.policy [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3baf998a00e494ba33dab6a0a2a88e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '836843da5be34c649d9a48a83e658288', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 983.702763] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 983.703298] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f06c313-92cf-4c76-bca6-ae263ee108f6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.712879] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 983.712879] env[61728]: value = "task-464625" [ 983.712879] env[61728]: _type = "Task" [ 983.712879] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.724156] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464625, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.985770] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464623, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450457} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.986269] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 705c8838-f605-46fa-b036-ec212f673f66/705c8838-f605-46fa-b036-ec212f673f66.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 983.986435] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 983.986994] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7d099b89-cb4a-4fe1-8478-1716acc491e5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.999394] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 983.999394] env[61728]: value = "task-464626" [ 983.999394] env[61728]: _type = "Task" [ 983.999394] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.012300] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464626, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.053122] env[61728]: DEBUG oslo_vmware.api [None req-22412f64-92b3-420b-9dd5-cd73d36cdb4b tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464624, 'name': PowerOnVM_Task} progress is 93%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.119914] env[61728]: DEBUG nova.compute.manager [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 984.231634] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464625, 'name': PowerOffVM_Task, 'duration_secs': 0.379206} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.234702] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 984.234983] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance 'f9f07464-9f43-43fd-8895-2673861747fa' progress to 17 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 984.338029] env[61728]: DEBUG nova.network.neutron [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Successfully created port: 18a00538-a062-4136-b2d4-2b3b3f783502 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 984.452871] env[61728]: DEBUG nova.objects.instance [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lazy-loading 'flavor' on Instance uuid 8c5b2380-bd14-451e-9612-2767c2a6adb2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.468901] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa4fc116-c79a-4f85-b172-18967b251ed0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.478239] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc8e13b-4ad3-46d1-9a54-873825e00794 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.515367] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb9cf38-480e-4064-b92e-1fb5dc63d49e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.526260] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e978e30-56e5-4744-bc6c-76a33a611aa9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.530201] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464626, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083672} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.530819] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 984.532014] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa91fdce-48d2-453f-b4bd-3ba4d4b81534 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.543243] env[61728]: DEBUG nova.compute.provider_tree [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.566403] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 705c8838-f605-46fa-b036-ec212f673f66/705c8838-f605-46fa-b036-ec212f673f66.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 984.568237] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b40884d-522d-4f68-8ecb-8e8d5727ce19 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.586162] env[61728]: DEBUG oslo_vmware.api [None req-22412f64-92b3-420b-9dd5-cd73d36cdb4b tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464624, 'name': PowerOnVM_Task, 'duration_secs': 0.677013} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.586827] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-22412f64-92b3-420b-9dd5-cd73d36cdb4b tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Resumed the VM {{(pid=61728) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 984.587073] env[61728]: DEBUG nova.compute.manager [None req-22412f64-92b3-420b-9dd5-cd73d36cdb4b tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 984.587852] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60e22c8-f6ef-4022-bd05-141ddb481836 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.592377] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 984.592377] env[61728]: value = "task-464627" [ 984.592377] env[61728]: _type = "Task" [ 984.592377] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.604911] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464627, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.741215] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 984.741497] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 984.741664] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.742176] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 984.742384] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.742721] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 984.743114] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 984.743371] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 984.743634] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 984.743817] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 984.744090] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.749078] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa8050fa-cc1e-4fed-914f-f1638ae1f223 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.765774] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 984.765774] env[61728]: value = "task-464628" [ 984.765774] env[61728]: _type = "Task" [ 984.765774] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.774456] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464628, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.958334] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.958562] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.958677] env[61728]: DEBUG nova.network.neutron [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 984.958875] env[61728]: DEBUG nova.objects.instance [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lazy-loading 'info_cache' on Instance uuid 8c5b2380-bd14-451e-9612-2767c2a6adb2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.057991] env[61728]: DEBUG nova.scheduler.client.report [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.106043] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464627, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.131347] env[61728]: DEBUG nova.compute.manager [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 985.161259] env[61728]: DEBUG nova.virt.hardware [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.161562] env[61728]: DEBUG nova.virt.hardware [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.161761] env[61728]: DEBUG nova.virt.hardware [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.161976] env[61728]: DEBUG nova.virt.hardware [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.162168] env[61728]: DEBUG nova.virt.hardware [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.162347] env[61728]: DEBUG nova.virt.hardware [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.162725] env[61728]: DEBUG nova.virt.hardware [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.162817] env[61728]: DEBUG nova.virt.hardware [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.162966] env[61728]: DEBUG nova.virt.hardware [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.163168] env[61728]: DEBUG nova.virt.hardware [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.163352] env[61728]: DEBUG nova.virt.hardware [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.164306] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b0173b-b6ea-4fc3-8579-9c93c0d19912 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.173248] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a8d2f9-9669-408d-a671-e46fb32ce0d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.276827] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464628, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.462318] env[61728]: DEBUG nova.objects.base [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Object Instance<8c5b2380-bd14-451e-9612-2767c2a6adb2> lazy-loaded attributes: flavor,info_cache {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 985.563722] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.564308] env[61728]: DEBUG nova.compute.manager [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 985.566953] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.677s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.567196] env[61728]: DEBUG nova.objects.instance [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lazy-loading 'resources' on Instance uuid ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.604778] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464627, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.780257] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464628, 'name': ReconfigVM_Task, 'duration_secs': 0.633604} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.780257] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance 'f9f07464-9f43-43fd-8895-2673861747fa' progress to 33 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 986.072151] env[61728]: DEBUG nova.compute.utils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 986.077024] env[61728]: DEBUG nova.compute.manager [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 986.077024] env[61728]: DEBUG nova.network.neutron [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 986.105861] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464627, 'name': ReconfigVM_Task, 'duration_secs': 1.458766} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.106508] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 705c8838-f605-46fa-b036-ec212f673f66/705c8838-f605-46fa-b036-ec212f673f66.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.113029] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-46809871-86b7-4046-88c2-e99f0318b69b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.121071] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 986.121071] env[61728]: value = "task-464629" [ 986.121071] env[61728]: _type = "Task" [ 986.121071] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.135311] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464629, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.179185] env[61728]: DEBUG nova.policy [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd32b5118a94042f797dea1fdf6bd05b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08367aaceba548fe93faaedf6371817d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 986.286505] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 986.286802] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 986.286947] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.287135] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 986.287286] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.287441] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 986.287654] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 986.287819] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 986.287994] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 986.288308] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 986.288618] env[61728]: DEBUG nova.virt.hardware [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 986.294960] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Reconfiguring VM instance instance-00000051 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 986.299008] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ee75c1f-336f-4e37-93d6-9adf3123a6f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.319022] env[61728]: DEBUG nova.compute.manager [req-e420abc7-bad0-4480-8151-6c1a9ecbb13d req-1c1664c5-3095-4e82-9f3e-37af2248a9b2 service nova] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Received event network-vif-plugged-18a00538-a062-4136-b2d4-2b3b3f783502 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 986.319022] env[61728]: DEBUG oslo_concurrency.lockutils [req-e420abc7-bad0-4480-8151-6c1a9ecbb13d req-1c1664c5-3095-4e82-9f3e-37af2248a9b2 service nova] Acquiring lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.319022] env[61728]: DEBUG oslo_concurrency.lockutils [req-e420abc7-bad0-4480-8151-6c1a9ecbb13d req-1c1664c5-3095-4e82-9f3e-37af2248a9b2 service nova] Lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.319022] env[61728]: DEBUG oslo_concurrency.lockutils [req-e420abc7-bad0-4480-8151-6c1a9ecbb13d req-1c1664c5-3095-4e82-9f3e-37af2248a9b2 service nova] Lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.319022] env[61728]: DEBUG nova.compute.manager [req-e420abc7-bad0-4480-8151-6c1a9ecbb13d req-1c1664c5-3095-4e82-9f3e-37af2248a9b2 service nova] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] No waiting events found dispatching network-vif-plugged-18a00538-a062-4136-b2d4-2b3b3f783502 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 986.319022] env[61728]: WARNING nova.compute.manager [req-e420abc7-bad0-4480-8151-6c1a9ecbb13d req-1c1664c5-3095-4e82-9f3e-37af2248a9b2 service nova] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Received unexpected event network-vif-plugged-18a00538-a062-4136-b2d4-2b3b3f783502 for instance with vm_state building and task_state spawning. [ 986.319022] env[61728]: DEBUG nova.network.neutron [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance_info_cache with network_info: [{"id": "4660e4da-30e8-4e51-9aea-a9375732537f", "address": "fa:16:3e:e1:e9:0c", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4660e4da-30", "ovs_interfaceid": "4660e4da-30e8-4e51-9aea-a9375732537f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.327117] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 986.327117] env[61728]: value = "task-464630" [ 986.327117] env[61728]: _type = "Task" [ 986.327117] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.341204] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464630, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.453867] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e819384-179f-4a44-8c79-e54a551e8946 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.463188] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53feb422-ea11-4f81-bd2c-4e086ab2647a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.470274] env[61728]: DEBUG nova.network.neutron [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Successfully updated port: 18a00538-a062-4136-b2d4-2b3b3f783502 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.501439] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "refresh_cache-0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.501439] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "refresh_cache-0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.501439] env[61728]: DEBUG nova.network.neutron [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 986.501439] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584cd8f0-5831-454a-a13d-b44f933ff03e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.512632] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd616864-3b08-415a-a26c-50d7239ee4a4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.535178] env[61728]: DEBUG nova.compute.provider_tree [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 986.537072] env[61728]: DEBUG nova.network.neutron [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Successfully created port: 03e4158f-eab0-4cd7-9d4e-b8ee77e099c6 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 986.577082] env[61728]: DEBUG nova.compute.manager [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 986.637620] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464629, 'name': Rename_Task, 'duration_secs': 0.217483} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.637620] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 986.637620] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0cbbe01-3b38-4648-b627-e17993c93921 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.643876] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 986.643876] env[61728]: value = "task-464631" [ 986.643876] env[61728]: _type = "Task" [ 986.643876] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.652793] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464631, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.820477] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.839458] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464630, 'name': ReconfigVM_Task, 'duration_secs': 0.222183} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.839768] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Reconfigured VM instance instance-00000051 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 986.840604] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a6c7be-17e6-4968-8daa-96e68f51f087 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.869152] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] f9f07464-9f43-43fd-8895-2673861747fa/f9f07464-9f43-43fd-8895-2673861747fa.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.869589] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b7b0b2f-9b4d-4657-91d3-03543590e7cb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.890033] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 986.890033] env[61728]: value = "task-464632" [ 986.890033] env[61728]: _type = "Task" [ 986.890033] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.898612] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464632, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.059729] env[61728]: DEBUG nova.network.neutron [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 987.088711] env[61728]: DEBUG nova.scheduler.client.report [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 150 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 987.088997] env[61728]: DEBUG nova.compute.provider_tree [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 150 to 151 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 987.089190] env[61728]: DEBUG nova.compute.provider_tree [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 987.156972] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464631, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.321704] env[61728]: DEBUG nova.network.neutron [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Updating instance_info_cache with network_info: [{"id": "18a00538-a062-4136-b2d4-2b3b3f783502", "address": "fa:16:3e:55:17:79", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18a00538-a0", "ovs_interfaceid": "18a00538-a062-4136-b2d4-2b3b3f783502", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.323708] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 987.324225] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4a6d4773-69b3-4300-a08d-336b569f19d9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.332475] env[61728]: DEBUG oslo_vmware.api [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 987.332475] env[61728]: value = "task-464633" [ 987.332475] env[61728]: _type = "Task" [ 987.332475] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.342257] env[61728]: DEBUG oslo_vmware.api [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464633, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.401791] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464632, 'name': ReconfigVM_Task, 'duration_secs': 0.363112} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.402185] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Reconfigured VM instance instance-00000051 to attach disk [datastore1] f9f07464-9f43-43fd-8895-2673861747fa/f9f07464-9f43-43fd-8895-2673861747fa.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.402453] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance 'f9f07464-9f43-43fd-8895-2673861747fa' progress to 50 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 987.594692] env[61728]: DEBUG nova.compute.manager [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 987.597478] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.030s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.599469] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.388s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.599727] env[61728]: DEBUG nova.objects.instance [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lazy-loading 'resources' on Instance uuid 2c166dd3-985c-4637-92ab-939b46a7c90e {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.616996] env[61728]: INFO nova.scheduler.client.report [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Deleted allocations for instance ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69 [ 987.624849] env[61728]: DEBUG nova.virt.hardware [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.625029] env[61728]: DEBUG nova.virt.hardware [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.625088] env[61728]: DEBUG nova.virt.hardware [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.625276] env[61728]: DEBUG nova.virt.hardware [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.626364] env[61728]: DEBUG nova.virt.hardware [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.626364] env[61728]: DEBUG nova.virt.hardware [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.626364] env[61728]: DEBUG nova.virt.hardware [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.626364] env[61728]: DEBUG nova.virt.hardware [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.626364] env[61728]: DEBUG nova.virt.hardware [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.626364] env[61728]: DEBUG nova.virt.hardware [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.626581] env[61728]: DEBUG nova.virt.hardware [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.631030] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7880842f-cfd7-46b2-b737-5abb4141b344 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.636722] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b14384-4bc5-49fd-830b-2110dbcb3830 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.659982] env[61728]: DEBUG oslo_vmware.api [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464631, 'name': PowerOnVM_Task, 'duration_secs': 0.51748} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.660294] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 987.660503] env[61728]: INFO nova.compute.manager [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Took 8.32 seconds to spawn the instance on the hypervisor. [ 987.660947] env[61728]: DEBUG nova.compute.manager [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 987.661478] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ea8980-307d-4c46-8ea4-8457ff81351a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.825116] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "refresh_cache-0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.825548] env[61728]: DEBUG nova.compute.manager [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Instance network_info: |[{"id": "18a00538-a062-4136-b2d4-2b3b3f783502", "address": "fa:16:3e:55:17:79", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18a00538-a0", "ovs_interfaceid": "18a00538-a062-4136-b2d4-2b3b3f783502", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 987.826033] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:17:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9732690c-bdcf-4e6f-9a32-42c196333eb8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '18a00538-a062-4136-b2d4-2b3b3f783502', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.837163] env[61728]: DEBUG oslo.service.loopingcall [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.837163] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 987.838866] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9df0c5b2-0f2f-4960-9a34-51bd4b3a4ca0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.859988] env[61728]: DEBUG oslo_vmware.api [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464633, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.861269] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.861269] env[61728]: value = "task-464634" [ 987.861269] env[61728]: _type = "Task" [ 987.861269] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.870118] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464634, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.910032] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aae2385-3905-46a7-8a63-f3b146b0e9dc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.933544] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc2a303-dc43-49f4-94d9-b88721a48ada {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.956435] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance 'f9f07464-9f43-43fd-8895-2673861747fa' progress to 67 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 988.036388] env[61728]: DEBUG nova.compute.manager [req-9500c763-de53-4e00-91f4-ca7a3897beca req-ee0ba0ca-e3bc-4d5f-bc7e-720ed6bc702d service nova] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Received event network-vif-plugged-03e4158f-eab0-4cd7-9d4e-b8ee77e099c6 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 988.036634] env[61728]: DEBUG oslo_concurrency.lockutils [req-9500c763-de53-4e00-91f4-ca7a3897beca req-ee0ba0ca-e3bc-4d5f-bc7e-720ed6bc702d service nova] Acquiring lock "172129ff-136e-4855-8659-3f1ac88dcbd2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.036867] env[61728]: DEBUG oslo_concurrency.lockutils [req-9500c763-de53-4e00-91f4-ca7a3897beca req-ee0ba0ca-e3bc-4d5f-bc7e-720ed6bc702d service nova] Lock "172129ff-136e-4855-8659-3f1ac88dcbd2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.039046] env[61728]: DEBUG oslo_concurrency.lockutils [req-9500c763-de53-4e00-91f4-ca7a3897beca req-ee0ba0ca-e3bc-4d5f-bc7e-720ed6bc702d service nova] Lock "172129ff-136e-4855-8659-3f1ac88dcbd2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.039046] env[61728]: DEBUG nova.compute.manager [req-9500c763-de53-4e00-91f4-ca7a3897beca req-ee0ba0ca-e3bc-4d5f-bc7e-720ed6bc702d service nova] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] No waiting events found dispatching network-vif-plugged-03e4158f-eab0-4cd7-9d4e-b8ee77e099c6 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 988.039046] env[61728]: WARNING nova.compute.manager [req-9500c763-de53-4e00-91f4-ca7a3897beca req-ee0ba0ca-e3bc-4d5f-bc7e-720ed6bc702d service nova] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Received unexpected event network-vif-plugged-03e4158f-eab0-4cd7-9d4e-b8ee77e099c6 for instance with vm_state building and task_state spawning. [ 988.041788] env[61728]: DEBUG nova.network.neutron [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Successfully updated port: 03e4158f-eab0-4cd7-9d4e-b8ee77e099c6 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 988.139640] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c03363a5-cbff-4748-9363-1e34aba60645 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.688s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.183283] env[61728]: INFO nova.compute.manager [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Took 30.71 seconds to build instance. [ 988.335759] env[61728]: DEBUG nova.compute.manager [req-98d68d1e-f4ce-47d6-8e95-1b8b6e860c96 req-3f1408f4-8d02-4dbb-a865-d07ff461a58c service nova] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Received event network-changed-18a00538-a062-4136-b2d4-2b3b3f783502 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 988.335986] env[61728]: DEBUG nova.compute.manager [req-98d68d1e-f4ce-47d6-8e95-1b8b6e860c96 req-3f1408f4-8d02-4dbb-a865-d07ff461a58c service nova] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Refreshing instance network info cache due to event network-changed-18a00538-a062-4136-b2d4-2b3b3f783502. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 988.336294] env[61728]: DEBUG oslo_concurrency.lockutils [req-98d68d1e-f4ce-47d6-8e95-1b8b6e860c96 req-3f1408f4-8d02-4dbb-a865-d07ff461a58c service nova] Acquiring lock "refresh_cache-0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.336438] env[61728]: DEBUG oslo_concurrency.lockutils [req-98d68d1e-f4ce-47d6-8e95-1b8b6e860c96 req-3f1408f4-8d02-4dbb-a865-d07ff461a58c service nova] Acquired lock "refresh_cache-0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.336611] env[61728]: DEBUG nova.network.neutron [req-98d68d1e-f4ce-47d6-8e95-1b8b6e860c96 req-3f1408f4-8d02-4dbb-a865-d07ff461a58c service nova] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Refreshing network info cache for port 18a00538-a062-4136-b2d4-2b3b3f783502 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 988.351742] env[61728]: DEBUG oslo_vmware.api [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464633, 'name': PowerOnVM_Task, 'duration_secs': 0.732957} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.351742] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 988.351909] env[61728]: DEBUG nova.compute.manager [None req-cb866832-e0b5-4653-b606-714b3c3df74c tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.352684] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0971f5ed-f2a8-4f70-9411-5723b39afc1e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.376596] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464634, 'name': CreateVM_Task, 'duration_secs': 0.363797} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.377866] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 988.377866] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.377866] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.378084] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 988.378482] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af274d24-4faf-4bcc-b744-53d516be8d4c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.385589] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 988.385589] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c0a9e4-2351-d10d-56c6-510aa0eb98d3" [ 988.385589] env[61728]: _type = "Task" [ 988.385589] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.397807] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c0a9e4-2351-d10d-56c6-510aa0eb98d3, 'name': SearchDatastore_Task, 'duration_secs': 0.010867} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.400945] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.401196] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.401431] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.401581] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.401808] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.403039] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e8271ac0-0e18-460f-b6a6-28298f0b9129 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.411818] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.412024] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 988.412805] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d80e71c-7972-4e85-a75b-739572b56112 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.420143] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 988.420143] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528cc661-487b-86f9-9e42-8876e8c99732" [ 988.420143] env[61728]: _type = "Task" [ 988.420143] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.429499] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528cc661-487b-86f9-9e42-8876e8c99732, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.441919] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d816e63-8fa3-4e46-8799-e3972d4423da {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.449596] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d28c0863-405f-45e7-b77c-78080b9e3dcc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.488272] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9143528f-c7d5-4184-bd51-a14716376346 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.502512] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3599c5-701c-4455-96bf-2c6ea577ae1f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.519821] env[61728]: DEBUG nova.compute.provider_tree [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.544451] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "refresh_cache-172129ff-136e-4855-8659-3f1ac88dcbd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.544624] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "refresh_cache-172129ff-136e-4855-8659-3f1ac88dcbd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.544869] env[61728]: DEBUG nova.network.neutron [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 988.684814] env[61728]: DEBUG oslo_concurrency.lockutils [None req-58fca204-d79a-447f-b5fd-b3f819872c21 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "705c8838-f605-46fa-b036-ec212f673f66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.223s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.933519] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528cc661-487b-86f9-9e42-8876e8c99732, 'name': SearchDatastore_Task, 'duration_secs': 0.009935} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.934516] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d444542b-d569-4664-a4da-42f57634acb8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.940370] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 988.940370] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520d96a6-8bcc-79a2-804a-1730cf2075c7" [ 988.940370] env[61728]: _type = "Task" [ 988.940370] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.949447] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520d96a6-8bcc-79a2-804a-1730cf2075c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.023646] env[61728]: DEBUG nova.scheduler.client.report [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 989.053091] env[61728]: DEBUG nova.network.neutron [req-98d68d1e-f4ce-47d6-8e95-1b8b6e860c96 req-3f1408f4-8d02-4dbb-a865-d07ff461a58c service nova] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Updated VIF entry in instance network info cache for port 18a00538-a062-4136-b2d4-2b3b3f783502. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 989.053474] env[61728]: DEBUG nova.network.neutron [req-98d68d1e-f4ce-47d6-8e95-1b8b6e860c96 req-3f1408f4-8d02-4dbb-a865-d07ff461a58c service nova] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Updating instance_info_cache with network_info: [{"id": "18a00538-a062-4136-b2d4-2b3b3f783502", "address": "fa:16:3e:55:17:79", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18a00538-a0", "ovs_interfaceid": "18a00538-a062-4136-b2d4-2b3b3f783502", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.109918] env[61728]: DEBUG nova.network.neutron [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 989.448551] env[61728]: DEBUG nova.network.neutron [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Updating instance_info_cache with network_info: [{"id": "03e4158f-eab0-4cd7-9d4e-b8ee77e099c6", "address": "fa:16:3e:3a:a5:a8", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03e4158f-ea", "ovs_interfaceid": "03e4158f-eab0-4cd7-9d4e-b8ee77e099c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.454967] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520d96a6-8bcc-79a2-804a-1730cf2075c7, 'name': SearchDatastore_Task, 'duration_secs': 0.010765} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.455584] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.456100] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f/0edd1f56-3e8e-4868-9c8e-a8caaf026e7f.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 989.456381] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-14b2f221-2f8c-468e-970a-4bb48f608afe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.465939] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 989.465939] env[61728]: value = "task-464635" [ 989.465939] env[61728]: _type = "Task" [ 989.465939] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.475467] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464635, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.529436] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.930s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.531868] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.897s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.534114] env[61728]: INFO nova.compute.claims [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 989.555784] env[61728]: INFO nova.scheduler.client.report [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Deleted allocations for instance 2c166dd3-985c-4637-92ab-939b46a7c90e [ 989.557412] env[61728]: DEBUG oslo_concurrency.lockutils [req-98d68d1e-f4ce-47d6-8e95-1b8b6e860c96 req-3f1408f4-8d02-4dbb-a865-d07ff461a58c service nova] Releasing lock "refresh_cache-0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.712972] env[61728]: DEBUG nova.network.neutron [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Port 83657468-5cac-4258-84f4-b105395d9cfa binding to destination host cpu-1 is already ACTIVE {{(pid=61728) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 989.959021] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "refresh_cache-172129ff-136e-4855-8659-3f1ac88dcbd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.959021] env[61728]: DEBUG nova.compute.manager [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Instance network_info: |[{"id": "03e4158f-eab0-4cd7-9d4e-b8ee77e099c6", "address": "fa:16:3e:3a:a5:a8", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03e4158f-ea", "ovs_interfaceid": "03e4158f-eab0-4cd7-9d4e-b8ee77e099c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 989.959021] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:a5:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd733acc2-07d0-479e-918c-ec8a21925389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03e4158f-eab0-4cd7-9d4e-b8ee77e099c6', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 989.980209] env[61728]: DEBUG oslo.service.loopingcall [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.981085] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 989.985497] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da85afbb-7fee-44f1-978e-5e5a9cde197d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.021852] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464635, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475164} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.023897] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f/0edd1f56-3e8e-4868-9c8e-a8caaf026e7f.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 990.024377] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 990.024747] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 990.024747] env[61728]: value = "task-464636" [ 990.024747] env[61728]: _type = "Task" [ 990.024747] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.025203] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5668018e-17fb-4da3-b2bc-34ffc39951ee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.037881] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464636, 'name': CreateVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.039839] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 990.039839] env[61728]: value = "task-464637" [ 990.039839] env[61728]: _type = "Task" [ 990.039839] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.055429] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464637, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.063520] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5e3d153-d5ff-45ee-aad9-89f9284606a3 tempest-ListServersNegativeTestJSON-702762766 tempest-ListServersNegativeTestJSON-702762766-project-member] Lock "2c166dd3-985c-4637-92ab-939b46a7c90e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.753s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.151833] env[61728]: DEBUG nova.compute.manager [req-a8d439bd-2825-47ce-9fc3-32f3b1f49532 req-91addcf4-a808-4eb9-874b-6f46f48d6c73 service nova] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Received event network-changed-03e4158f-eab0-4cd7-9d4e-b8ee77e099c6 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 990.152027] env[61728]: DEBUG nova.compute.manager [req-a8d439bd-2825-47ce-9fc3-32f3b1f49532 req-91addcf4-a808-4eb9-874b-6f46f48d6c73 service nova] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Refreshing instance network info cache due to event network-changed-03e4158f-eab0-4cd7-9d4e-b8ee77e099c6. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 990.152529] env[61728]: DEBUG oslo_concurrency.lockutils [req-a8d439bd-2825-47ce-9fc3-32f3b1f49532 req-91addcf4-a808-4eb9-874b-6f46f48d6c73 service nova] Acquiring lock "refresh_cache-172129ff-136e-4855-8659-3f1ac88dcbd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.152823] env[61728]: DEBUG oslo_concurrency.lockutils [req-a8d439bd-2825-47ce-9fc3-32f3b1f49532 req-91addcf4-a808-4eb9-874b-6f46f48d6c73 service nova] Acquired lock "refresh_cache-172129ff-136e-4855-8659-3f1ac88dcbd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.153079] env[61728]: DEBUG nova.network.neutron [req-a8d439bd-2825-47ce-9fc3-32f3b1f49532 req-91addcf4-a808-4eb9-874b-6f46f48d6c73 service nova] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Refreshing network info cache for port 03e4158f-eab0-4cd7-9d4e-b8ee77e099c6 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 990.430351] env[61728]: DEBUG nova.compute.manager [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Received event network-changed-c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 990.430562] env[61728]: DEBUG nova.compute.manager [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Refreshing instance network info cache due to event network-changed-c8407eac-4337-4ae3-a071-08f48bf87248. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 990.430795] env[61728]: DEBUG oslo_concurrency.lockutils [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] Acquiring lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.430983] env[61728]: DEBUG oslo_concurrency.lockutils [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] Acquired lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.431994] env[61728]: DEBUG nova.network.neutron [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Refreshing network info cache for port c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 990.540306] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464636, 'name': CreateVM_Task, 'duration_secs': 0.353687} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.540306] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 990.540306] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.540306] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.540306] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 990.540629] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce3fc6ea-8e5d-4a0c-8eb5-f4ad00a2c2fb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.554869] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464637, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072078} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.555137] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 990.555137] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529ecff2-680d-563c-3866-7fd913b03c32" [ 990.555137] env[61728]: _type = "Task" [ 990.555137] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.555405] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 990.560494] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adaeefbc-4978-4152-98ef-0a203c2d365a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.598098] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f/0edd1f56-3e8e-4868-9c8e-a8caaf026e7f.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 990.604869] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd0d6fe1-a903-4321-89a8-ebd6514a56e2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.620264] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529ecff2-680d-563c-3866-7fd913b03c32, 'name': SearchDatastore_Task, 'duration_secs': 0.010391} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.620264] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.620568] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.620850] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.620850] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.621161] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.621728] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-24bc45bb-7ee9-4f0a-a68c-591b7de6e943 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.630795] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 990.630795] env[61728]: value = "task-464638" [ 990.630795] env[61728]: _type = "Task" [ 990.630795] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.634319] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.634512] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 990.635869] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4bf62b6f-3ae1-48fb-8126-da8e69ecffb3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.643691] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464638, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.647796] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 990.647796] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e9bd79-b0c8-320c-5f24-35cf78d21cfa" [ 990.647796] env[61728]: _type = "Task" [ 990.647796] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.658704] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e9bd79-b0c8-320c-5f24-35cf78d21cfa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.751481] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "f9f07464-9f43-43fd-8895-2673861747fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.751481] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "f9f07464-9f43-43fd-8895-2673861747fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.751481] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "f9f07464-9f43-43fd-8895-2673861747fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.952016] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f80484d3-6b09-46b3-af4c-a2219beef8bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.961775] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f9c866-6ffa-4511-881c-6d9ddd831608 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.000963] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db6602ee-ab91-44a9-b256-e19002b54e3b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.010152] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79cab7d7-e772-4b81-9bfe-1a6263263466 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.027971] env[61728]: DEBUG nova.compute.provider_tree [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.141893] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464638, 'name': ReconfigVM_Task, 'duration_secs': 0.287063} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.141893] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f/0edd1f56-3e8e-4868-9c8e-a8caaf026e7f.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 991.141893] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0820d768-6a7d-444f-a0bb-ae0d3d7b23a3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.152925] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 991.152925] env[61728]: value = "task-464639" [ 991.152925] env[61728]: _type = "Task" [ 991.152925] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.160498] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e9bd79-b0c8-320c-5f24-35cf78d21cfa, 'name': SearchDatastore_Task, 'duration_secs': 0.013801} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.165868] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccb5c806-3f00-4e61-be72-977d4f931906 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.171908] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464639, 'name': Rename_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.176082] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 991.176082] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5261ba09-9fae-7cb6-711b-a5ad2772cd98" [ 991.176082] env[61728]: _type = "Task" [ 991.176082] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.186809] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5261ba09-9fae-7cb6-711b-a5ad2772cd98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.344437] env[61728]: DEBUG nova.network.neutron [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updated VIF entry in instance network info cache for port c8407eac-4337-4ae3-a071-08f48bf87248. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 991.344838] env[61728]: DEBUG nova.network.neutron [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updating instance_info_cache with network_info: [{"id": "c8407eac-4337-4ae3-a071-08f48bf87248", "address": "fa:16:3e:2d:a6:d6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8407eac-43", "ovs_interfaceid": "c8407eac-4337-4ae3-a071-08f48bf87248", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.390500] env[61728]: DEBUG nova.network.neutron [req-a8d439bd-2825-47ce-9fc3-32f3b1f49532 req-91addcf4-a808-4eb9-874b-6f46f48d6c73 service nova] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Updated VIF entry in instance network info cache for port 03e4158f-eab0-4cd7-9d4e-b8ee77e099c6. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 991.391216] env[61728]: DEBUG nova.network.neutron [req-a8d439bd-2825-47ce-9fc3-32f3b1f49532 req-91addcf4-a808-4eb9-874b-6f46f48d6c73 service nova] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Updating instance_info_cache with network_info: [{"id": "03e4158f-eab0-4cd7-9d4e-b8ee77e099c6", "address": "fa:16:3e:3a:a5:a8", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03e4158f-ea", "ovs_interfaceid": "03e4158f-eab0-4cd7-9d4e-b8ee77e099c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.531316] env[61728]: DEBUG nova.scheduler.client.report [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 991.563400] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.563897] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.563979] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.564201] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.564382] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.566400] env[61728]: INFO nova.compute.manager [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Terminating instance [ 991.568123] env[61728]: DEBUG nova.compute.manager [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 991.568321] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 991.570058] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc45232-975a-402b-9217-7153bc31c870 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.578184] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 991.578617] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b6553aca-4690-4cf1-bf81-a5c24f492c01 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.587360] env[61728]: DEBUG oslo_vmware.api [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 991.587360] env[61728]: value = "task-464640" [ 991.587360] env[61728]: _type = "Task" [ 991.587360] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.601926] env[61728]: DEBUG oslo_vmware.api [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464640, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.662894] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464639, 'name': Rename_Task, 'duration_secs': 0.169029} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.663189] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 991.663465] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a40c574e-e7ff-484f-a094-6a10eaa8cac5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.670338] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 991.670338] env[61728]: value = "task-464641" [ 991.670338] env[61728]: _type = "Task" [ 991.670338] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.678645] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464641, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.689774] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5261ba09-9fae-7cb6-711b-a5ad2772cd98, 'name': SearchDatastore_Task, 'duration_secs': 0.021698} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.690055] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.690328] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 172129ff-136e-4855-8659-3f1ac88dcbd2/172129ff-136e-4855-8659-3f1ac88dcbd2.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 991.690647] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1dc7135a-c762-4a20-8107-ae2ec21990b2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.698109] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 991.698109] env[61728]: value = "task-464642" [ 991.698109] env[61728]: _type = "Task" [ 991.698109] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.708386] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464642, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.824377] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.826287] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.826287] env[61728]: DEBUG nova.network.neutron [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 991.847358] env[61728]: DEBUG oslo_concurrency.lockutils [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] Releasing lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.847486] env[61728]: DEBUG nova.compute.manager [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Received event network-changed-ec0ae08d-62b0-4506-bc0b-65e9152dd5e8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 991.847621] env[61728]: DEBUG nova.compute.manager [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Refreshing instance network info cache due to event network-changed-ec0ae08d-62b0-4506-bc0b-65e9152dd5e8. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 991.847879] env[61728]: DEBUG oslo_concurrency.lockutils [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] Acquiring lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.848078] env[61728]: DEBUG oslo_concurrency.lockutils [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] Acquired lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.848274] env[61728]: DEBUG nova.network.neutron [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Refreshing network info cache for port ec0ae08d-62b0-4506-bc0b-65e9152dd5e8 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 991.894652] env[61728]: DEBUG oslo_concurrency.lockutils [req-a8d439bd-2825-47ce-9fc3-32f3b1f49532 req-91addcf4-a808-4eb9-874b-6f46f48d6c73 service nova] Releasing lock "refresh_cache-172129ff-136e-4855-8659-3f1ac88dcbd2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.036722] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.036971] env[61728]: DEBUG nova.compute.manager [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 992.039706] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.859s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.041285] env[61728]: INFO nova.compute.claims [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 992.104999] env[61728]: DEBUG oslo_vmware.api [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464640, 'name': PowerOffVM_Task, 'duration_secs': 0.330568} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.105303] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 992.105477] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 992.105748] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-647110a4-e857-4d94-847a-486da60784bf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.182277] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464641, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.183650] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 992.184236] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 992.184236] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Deleting the datastore file [datastore1] 3bfefcfc-db97-4a9d-86cb-9fb1d8158863 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 992.184366] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-35c7b233-07c2-471b-918a-0674f30e79ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.190831] env[61728]: DEBUG oslo_vmware.api [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for the task: (returnval){ [ 992.190831] env[61728]: value = "task-464644" [ 992.190831] env[61728]: _type = "Task" [ 992.190831] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.196513] env[61728]: DEBUG nova.compute.manager [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Received event network-changed-ec0ae08d-62b0-4506-bc0b-65e9152dd5e8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 992.196701] env[61728]: DEBUG nova.compute.manager [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Refreshing instance network info cache due to event network-changed-ec0ae08d-62b0-4506-bc0b-65e9152dd5e8. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 992.196895] env[61728]: DEBUG oslo_concurrency.lockutils [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] Acquiring lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.205053] env[61728]: DEBUG oslo_vmware.api [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464644, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.210906] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464642, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475031} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.211179] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 172129ff-136e-4855-8659-3f1ac88dcbd2/172129ff-136e-4855-8659-3f1ac88dcbd2.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 992.211401] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 992.211654] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5484609d-0c7d-4a98-8496-46f10ab55c26 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.220376] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 992.220376] env[61728]: value = "task-464645" [ 992.220376] env[61728]: _type = "Task" [ 992.220376] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.229505] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464645, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.548465] env[61728]: DEBUG nova.compute.utils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 992.552353] env[61728]: DEBUG nova.compute.manager [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 992.552538] env[61728]: DEBUG nova.network.neutron [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 992.635613] env[61728]: DEBUG nova.policy [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54c30aeeea3a4c1b8ed34430361c2a4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a02bd852c564acd99e12fd17279f028', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 992.677221] env[61728]: DEBUG nova.network.neutron [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance_info_cache with network_info: [{"id": "83657468-5cac-4258-84f4-b105395d9cfa", "address": "fa:16:3e:55:1a:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83657468-5c", "ovs_interfaceid": "83657468-5cac-4258-84f4-b105395d9cfa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.685668] env[61728]: DEBUG oslo_vmware.api [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464641, 'name': PowerOnVM_Task, 'duration_secs': 0.69084} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.686311] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 992.686311] env[61728]: INFO nova.compute.manager [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Took 7.55 seconds to spawn the instance on the hypervisor. [ 992.686436] env[61728]: DEBUG nova.compute.manager [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 992.687300] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0caf2326-dea9-4ce7-a58e-dfbed4a6aca6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.706391] env[61728]: DEBUG oslo_vmware.api [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Task: {'id': task-464644, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171916} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.706650] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.706832] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 992.707054] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 992.707247] env[61728]: INFO nova.compute.manager [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Took 1.14 seconds to destroy the instance on the hypervisor. [ 992.707494] env[61728]: DEBUG oslo.service.loopingcall [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.707682] env[61728]: DEBUG nova.compute.manager [-] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 992.708242] env[61728]: DEBUG nova.network.neutron [-] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 992.726739] env[61728]: DEBUG nova.network.neutron [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Updated VIF entry in instance network info cache for port ec0ae08d-62b0-4506-bc0b-65e9152dd5e8. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 992.727478] env[61728]: DEBUG nova.network.neutron [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Updating instance_info_cache with network_info: [{"id": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "address": "fa:16:3e:57:58:2b", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec0ae08d-62", "ovs_interfaceid": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.731943] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464645, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070276} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.732421] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 992.733293] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de05a11-e26b-4acc-b009-f64c7661903d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.760036] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 172129ff-136e-4855-8659-3f1ac88dcbd2/172129ff-136e-4855-8659-3f1ac88dcbd2.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.763565] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7977d0fb-332a-4528-a5e8-2f5b8c22f7e1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.786056] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 992.786056] env[61728]: value = "task-464646" [ 992.786056] env[61728]: _type = "Task" [ 992.786056] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.795351] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464646, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.997610] env[61728]: DEBUG nova.network.neutron [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Successfully created port: 81ffb8f2-c59b-4b0e-894a-d8633440d6eb {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 993.057023] env[61728]: DEBUG nova.compute.manager [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 993.180485] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.213035] env[61728]: INFO nova.compute.manager [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Took 34.48 seconds to build instance. [ 993.220876] env[61728]: DEBUG nova.compute.manager [req-40ce6b7d-1406-4836-86c8-0e77224cc694 req-3c604d06-25e0-44c1-a421-6cd176170c71 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Received event network-vif-deleted-124d4780-15ec-4f17-a2c4-edf1517ed589 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 993.221193] env[61728]: INFO nova.compute.manager [req-40ce6b7d-1406-4836-86c8-0e77224cc694 req-3c604d06-25e0-44c1-a421-6cd176170c71 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Neutron deleted interface 124d4780-15ec-4f17-a2c4-edf1517ed589; detaching it from the instance and deleting it from the info cache [ 993.221269] env[61728]: DEBUG nova.network.neutron [req-40ce6b7d-1406-4836-86c8-0e77224cc694 req-3c604d06-25e0-44c1-a421-6cd176170c71 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.233327] env[61728]: DEBUG oslo_concurrency.lockutils [req-081c2d00-501d-4614-8be9-77f792e51126 req-fe875689-c231-4f9c-a11c-1036a3587e1a service nova] Releasing lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.234946] env[61728]: DEBUG oslo_concurrency.lockutils [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] Acquired lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.234946] env[61728]: DEBUG nova.network.neutron [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Refreshing network info cache for port ec0ae08d-62b0-4506-bc0b-65e9152dd5e8 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 993.300630] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464646, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.414184] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f96af6-6d04-4c32-be78-28a79fdbf0e0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.423222] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2845b8-1731-44bb-8856-28242c93b843 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.457550] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f100ea00-6440-4deb-9d73-0f65ce101d4d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.467764] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d97742-956d-40cf-848c-f2102178ae76 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.485691] env[61728]: DEBUG nova.compute.provider_tree [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.628175] env[61728]: DEBUG nova.network.neutron [-] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.690649] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b5c7c7-4bcd-4e7e-9629-610e82f32b98 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.698688] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee155de-4eef-4ff0-bca6-eb930afe751c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.709069] env[61728]: DEBUG oslo_concurrency.lockutils [None req-41607e8c-26ff-4378-b80d-3626678afbd5 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.715311] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4bda65fe-97cc-41d4-ace8-5acd9f498787 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.998s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.715578] env[61728]: DEBUG oslo_concurrency.lockutils [None req-41607e8c-26ff-4378-b80d-3626678afbd5 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.007s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.715723] env[61728]: DEBUG nova.compute.manager [None req-41607e8c-26ff-4378-b80d-3626678afbd5 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 993.716502] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1930fe4d-a175-404d-9e90-08e2aa4afc31 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.723091] env[61728]: DEBUG nova.compute.manager [None req-41607e8c-26ff-4378-b80d-3626678afbd5 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61728) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 993.723666] env[61728]: DEBUG nova.objects.instance [None req-41607e8c-26ff-4378-b80d-3626678afbd5 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lazy-loading 'flavor' on Instance uuid 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.729080] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db4b5600-e78d-4d1d-86e8-3c336649db6a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.743586] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac33bd3d-2b44-40d3-aadd-a7b654c4226d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.500095] env[61728]: DEBUG nova.scheduler.client.report [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 994.503938] env[61728]: DEBUG nova.compute.manager [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 994.506085] env[61728]: INFO nova.compute.manager [-] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Took 1.80 seconds to deallocate network for instance. [ 994.508191] env[61728]: DEBUG nova.compute.manager [req-40ce6b7d-1406-4836-86c8-0e77224cc694 req-3c604d06-25e0-44c1-a421-6cd176170c71 service nova] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Detach interface failed, port_id=124d4780-15ec-4f17-a2c4-edf1517ed589, reason: Instance 3bfefcfc-db97-4a9d-86cb-9fb1d8158863 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 994.517828] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-41607e8c-26ff-4378-b80d-3626678afbd5 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 994.519891] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9fdd5677-4771-4982-9383-72a54a42d6d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.530327] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464646, 'name': ReconfigVM_Task, 'duration_secs': 0.753155} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.533695] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 172129ff-136e-4855-8659-3f1ac88dcbd2/172129ff-136e-4855-8659-3f1ac88dcbd2.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 994.534822] env[61728]: DEBUG oslo_vmware.api [None req-41607e8c-26ff-4378-b80d-3626678afbd5 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 994.534822] env[61728]: value = "task-464647" [ 994.534822] env[61728]: _type = "Task" [ 994.534822] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.535941] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39be501f-505f-43cd-a1d1-ea2f9d77ce0f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.548069] env[61728]: DEBUG oslo_vmware.api [None req-41607e8c-26ff-4378-b80d-3626678afbd5 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464647, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.550502] env[61728]: DEBUG nova.virt.hardware [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 994.550777] env[61728]: DEBUG nova.virt.hardware [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 994.550936] env[61728]: DEBUG nova.virt.hardware [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 994.551139] env[61728]: DEBUG nova.virt.hardware [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 994.551287] env[61728]: DEBUG nova.virt.hardware [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 994.551434] env[61728]: DEBUG nova.virt.hardware [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 994.551644] env[61728]: DEBUG nova.virt.hardware [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 994.551847] env[61728]: DEBUG nova.virt.hardware [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 994.552024] env[61728]: DEBUG nova.virt.hardware [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 994.552196] env[61728]: DEBUG nova.virt.hardware [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 994.552627] env[61728]: DEBUG nova.virt.hardware [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 994.554412] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d31e49c-0940-411d-8587-167e688c729c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.557681] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 994.557681] env[61728]: value = "task-464648" [ 994.557681] env[61728]: _type = "Task" [ 994.557681] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.567036] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c852c5c2-31a1-4862-9f20-547da9f9b22d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.575591] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464648, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.588336] env[61728]: DEBUG nova.compute.manager [req-074a0c0e-0a23-40c3-abca-0be5254ea4c9 req-d995282c-7439-4706-926c-6702f06d4699 service nova] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Received event network-vif-plugged-81ffb8f2-c59b-4b0e-894a-d8633440d6eb {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 994.588545] env[61728]: DEBUG oslo_concurrency.lockutils [req-074a0c0e-0a23-40c3-abca-0be5254ea4c9 req-d995282c-7439-4706-926c-6702f06d4699 service nova] Acquiring lock "b4241d6f-1d70-4f61-a322-caff7c4e6f64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.589709] env[61728]: DEBUG oslo_concurrency.lockutils [req-074a0c0e-0a23-40c3-abca-0be5254ea4c9 req-d995282c-7439-4706-926c-6702f06d4699 service nova] Lock "b4241d6f-1d70-4f61-a322-caff7c4e6f64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.589709] env[61728]: DEBUG oslo_concurrency.lockutils [req-074a0c0e-0a23-40c3-abca-0be5254ea4c9 req-d995282c-7439-4706-926c-6702f06d4699 service nova] Lock "b4241d6f-1d70-4f61-a322-caff7c4e6f64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.589709] env[61728]: DEBUG nova.compute.manager [req-074a0c0e-0a23-40c3-abca-0be5254ea4c9 req-d995282c-7439-4706-926c-6702f06d4699 service nova] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] No waiting events found dispatching network-vif-plugged-81ffb8f2-c59b-4b0e-894a-d8633440d6eb {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 994.589709] env[61728]: WARNING nova.compute.manager [req-074a0c0e-0a23-40c3-abca-0be5254ea4c9 req-d995282c-7439-4706-926c-6702f06d4699 service nova] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Received unexpected event network-vif-plugged-81ffb8f2-c59b-4b0e-894a-d8633440d6eb for instance with vm_state building and task_state spawning. [ 994.624996] env[61728]: DEBUG nova.network.neutron [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Successfully updated port: 81ffb8f2-c59b-4b0e-894a-d8633440d6eb {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 994.794528] env[61728]: DEBUG nova.network.neutron [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Updated VIF entry in instance network info cache for port ec0ae08d-62b0-4506-bc0b-65e9152dd5e8. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 994.794917] env[61728]: DEBUG nova.network.neutron [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Updating instance_info_cache with network_info: [{"id": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "address": "fa:16:3e:57:58:2b", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec0ae08d-62", "ovs_interfaceid": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.013058] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.973s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.013383] env[61728]: DEBUG nova.compute.manager [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 995.016385] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.988s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.016601] env[61728]: DEBUG nova.objects.instance [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lazy-loading 'resources' on Instance uuid 0af88a53-bbe7-443e-a7dc-734e483bb730 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.025941] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.051211] env[61728]: DEBUG oslo_vmware.api [None req-41607e8c-26ff-4378-b80d-3626678afbd5 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464647, 'name': PowerOffVM_Task, 'duration_secs': 0.196738} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.051519] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-41607e8c-26ff-4378-b80d-3626678afbd5 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 995.051733] env[61728]: DEBUG nova.compute.manager [None req-41607e8c-26ff-4378-b80d-3626678afbd5 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.052782] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11cdfc62-c1b7-4a26-ab5b-2cd2afb18dae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.075622] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464648, 'name': Rename_Task, 'duration_secs': 0.163994} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.076315] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 995.076590] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eb6ce340-929a-4047-9658-5815e8e6d21c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.084555] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b11970-bcef-41ec-88cd-1c0aa1357e7a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.103337] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 995.103337] env[61728]: value = "task-464649" [ 995.103337] env[61728]: _type = "Task" [ 995.103337] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.124910] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2034b6-ff76-4bf0-a99e-40b5daf14a68 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.127620] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "refresh_cache-b4241d6f-1d70-4f61-a322-caff7c4e6f64" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.127734] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "refresh_cache-b4241d6f-1d70-4f61-a322-caff7c4e6f64" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.127877] env[61728]: DEBUG nova.network.neutron [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 995.135182] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance 'f9f07464-9f43-43fd-8895-2673861747fa' progress to 83 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 995.145450] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464649, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.298150] env[61728]: DEBUG oslo_concurrency.lockutils [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] Releasing lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.298584] env[61728]: DEBUG nova.compute.manager [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Received event network-changed-c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 995.298678] env[61728]: DEBUG nova.compute.manager [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Refreshing instance network info cache due to event network-changed-c8407eac-4337-4ae3-a071-08f48bf87248. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 995.298857] env[61728]: DEBUG oslo_concurrency.lockutils [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] Acquiring lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.299012] env[61728]: DEBUG oslo_concurrency.lockutils [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] Acquired lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.299190] env[61728]: DEBUG nova.network.neutron [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Refreshing network info cache for port c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 995.519913] env[61728]: DEBUG nova.compute.utils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 995.524324] env[61728]: DEBUG nova.compute.manager [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 995.524727] env[61728]: DEBUG nova.network.neutron [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 995.565375] env[61728]: DEBUG oslo_concurrency.lockutils [None req-41607e8c-26ff-4378-b80d-3626678afbd5 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.849s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.569298] env[61728]: DEBUG nova.policy [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '70856f8255c940ac979244efb02c43b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '276942c8cf8a42729a541de096f69a0c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 995.629742] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464649, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.643631] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 995.646358] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b6b31c7-0303-470a-af6d-7b0310e88453 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.655605] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 995.655605] env[61728]: value = "task-464650" [ 995.655605] env[61728]: _type = "Task" [ 995.655605] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.662785] env[61728]: DEBUG nova.network.neutron [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 995.667462] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464650, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.826037] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e7177e-c67a-48e7-abd8-33ab13d9f6f0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.830801] env[61728]: DEBUG nova.network.neutron [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Updating instance_info_cache with network_info: [{"id": "81ffb8f2-c59b-4b0e-894a-d8633440d6eb", "address": "fa:16:3e:cb:69:dc", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81ffb8f2-c5", "ovs_interfaceid": "81ffb8f2-c59b-4b0e-894a-d8633440d6eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.835977] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caec5fe6-e620-4eea-abb9-ab0c116431dc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.872710] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-734178ec-3d02-46ac-b68c-8fe4381cd285 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.881941] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0acbb7d2-474f-4148-8136-910f2bca302f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.897042] env[61728]: DEBUG nova.compute.provider_tree [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.030260] env[61728]: DEBUG nova.compute.manager [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 996.097165] env[61728]: DEBUG nova.network.neutron [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Successfully created port: eb281159-3780-4ea6-b463-cd8d839be907 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 996.102867] env[61728]: DEBUG nova.network.neutron [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updated VIF entry in instance network info cache for port c8407eac-4337-4ae3-a071-08f48bf87248. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 996.106478] env[61728]: DEBUG nova.network.neutron [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updating instance_info_cache with network_info: [{"id": "c8407eac-4337-4ae3-a071-08f48bf87248", "address": "fa:16:3e:2d:a6:d6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8407eac-43", "ovs_interfaceid": "c8407eac-4337-4ae3-a071-08f48bf87248", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.130418] env[61728]: DEBUG oslo_vmware.api [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464649, 'name': PowerOnVM_Task, 'duration_secs': 0.956516} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.130788] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 996.131047] env[61728]: INFO nova.compute.manager [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Took 8.54 seconds to spawn the instance on the hypervisor. [ 996.131233] env[61728]: DEBUG nova.compute.manager [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 996.132270] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6fdac8-2475-4aa3-870e-2ce2005006dd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.171113] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464650, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.336116] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "refresh_cache-b4241d6f-1d70-4f61-a322-caff7c4e6f64" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.336481] env[61728]: DEBUG nova.compute.manager [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Instance network_info: |[{"id": "81ffb8f2-c59b-4b0e-894a-d8633440d6eb", "address": "fa:16:3e:cb:69:dc", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81ffb8f2-c5", "ovs_interfaceid": "81ffb8f2-c59b-4b0e-894a-d8633440d6eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 996.336939] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:69:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc6d5964-1106-4345-a26d-185dabd4ff0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '81ffb8f2-c59b-4b0e-894a-d8633440d6eb', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 996.345293] env[61728]: DEBUG oslo.service.loopingcall [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.345622] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 996.346238] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0dd628eb-913c-4560-b2f2-00f75d0eef01 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.368983] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 996.368983] env[61728]: value = "task-464651" [ 996.368983] env[61728]: _type = "Task" [ 996.368983] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.377340] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464651, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.400752] env[61728]: DEBUG nova.scheduler.client.report [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.530561] env[61728]: INFO nova.compute.manager [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Rebuilding instance [ 996.589260] env[61728]: DEBUG nova.compute.manager [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 996.590510] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5ffece-7ada-4518-a5ab-490855a27f51 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.605970] env[61728]: DEBUG oslo_concurrency.lockutils [req-8e7f977e-3bd3-45f7-9531-df8d86e2e5ff req-df6dd522-d126-43fe-a743-f60d14b617f1 service nova] Releasing lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.611024] env[61728]: DEBUG nova.compute.manager [req-d0f7e9b6-1828-4cdf-b5ef-643e945692c3 req-1e2b1f71-b445-4c5d-a0c8-935f00452ce3 service nova] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Received event network-changed-81ffb8f2-c59b-4b0e-894a-d8633440d6eb {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 996.611024] env[61728]: DEBUG nova.compute.manager [req-d0f7e9b6-1828-4cdf-b5ef-643e945692c3 req-1e2b1f71-b445-4c5d-a0c8-935f00452ce3 service nova] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Refreshing instance network info cache due to event network-changed-81ffb8f2-c59b-4b0e-894a-d8633440d6eb. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 996.611024] env[61728]: DEBUG oslo_concurrency.lockutils [req-d0f7e9b6-1828-4cdf-b5ef-643e945692c3 req-1e2b1f71-b445-4c5d-a0c8-935f00452ce3 service nova] Acquiring lock "refresh_cache-b4241d6f-1d70-4f61-a322-caff7c4e6f64" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.611418] env[61728]: DEBUG oslo_concurrency.lockutils [req-d0f7e9b6-1828-4cdf-b5ef-643e945692c3 req-1e2b1f71-b445-4c5d-a0c8-935f00452ce3 service nova] Acquired lock "refresh_cache-b4241d6f-1d70-4f61-a322-caff7c4e6f64" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.611418] env[61728]: DEBUG nova.network.neutron [req-d0f7e9b6-1828-4cdf-b5ef-643e945692c3 req-1e2b1f71-b445-4c5d-a0c8-935f00452ce3 service nova] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Refreshing network info cache for port 81ffb8f2-c59b-4b0e-894a-d8633440d6eb {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 996.650767] env[61728]: INFO nova.compute.manager [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Took 37.08 seconds to build instance. [ 996.668617] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464650, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.879098] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464651, 'name': CreateVM_Task, 'duration_secs': 0.382608} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.879329] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 996.879998] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.880195] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.880530] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 996.880791] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c509b4ec-8cfd-4b3d-b131-7a4e67e221d1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.886374] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 996.886374] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52548d9c-0f34-d719-2d6a-e131191fdb55" [ 996.886374] env[61728]: _type = "Task" [ 996.886374] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.895301] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52548d9c-0f34-d719-2d6a-e131191fdb55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.906854] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.890s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.909125] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.499s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.911299] env[61728]: INFO nova.compute.claims [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 996.928892] env[61728]: INFO nova.scheduler.client.report [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Deleted allocations for instance 0af88a53-bbe7-443e-a7dc-734e483bb730 [ 996.952140] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b395b69-48e6-46a9-bc61-0b00f8b18e46 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.961291] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-12fcc626-3d68-414c-9d1a-78999e439a45 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Suspending the VM {{(pid=61728) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 996.961541] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-818cda18-b919-4998-a5ae-78bc1424f127 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.968942] env[61728]: DEBUG oslo_vmware.api [None req-12fcc626-3d68-414c-9d1a-78999e439a45 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 996.968942] env[61728]: value = "task-464652" [ 996.968942] env[61728]: _type = "Task" [ 996.968942] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.978879] env[61728]: DEBUG oslo_vmware.api [None req-12fcc626-3d68-414c-9d1a-78999e439a45 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464652, 'name': SuspendVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.043020] env[61728]: DEBUG nova.compute.manager [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 997.071387] env[61728]: DEBUG nova.virt.hardware [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 997.071665] env[61728]: DEBUG nova.virt.hardware [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 997.071908] env[61728]: DEBUG nova.virt.hardware [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 997.072129] env[61728]: DEBUG nova.virt.hardware [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 997.072329] env[61728]: DEBUG nova.virt.hardware [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 997.072593] env[61728]: DEBUG nova.virt.hardware [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 997.072877] env[61728]: DEBUG nova.virt.hardware [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 997.073323] env[61728]: DEBUG nova.virt.hardware [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 997.073524] env[61728]: DEBUG nova.virt.hardware [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 997.073701] env[61728]: DEBUG nova.virt.hardware [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 997.073892] env[61728]: DEBUG nova.virt.hardware [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 997.074802] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496a69da-f70b-4117-a8d8-213ac35987c0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.088462] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66fda6b-2f9b-4e8e-adf3-5520ca99417b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.106328] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 997.106861] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-073aa6f3-2015-4b63-a350-e3441424fb70 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.116439] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 997.116439] env[61728]: value = "task-464653" [ 997.116439] env[61728]: _type = "Task" [ 997.116439] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.128156] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 997.128415] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 997.129748] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c2c2d6-e22b-4ce9-8c61-abc370a0a1ee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.139038] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 997.139327] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af1d34a8-6a0b-44ce-9ee7-bb6457d77628 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.153343] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8402232b-b8ef-4cb3-8fb0-3ae934294ce7 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "172129ff-136e-4855-8659-3f1ac88dcbd2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.587s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.168131] env[61728]: DEBUG oslo_vmware.api [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464650, 'name': PowerOnVM_Task, 'duration_secs': 1.445548} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.168419] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 997.168703] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4807872c-7db4-4126-b202-f48a4653f080 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance 'f9f07464-9f43-43fd-8895-2673861747fa' progress to 100 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 997.346578] env[61728]: DEBUG nova.network.neutron [req-d0f7e9b6-1828-4cdf-b5ef-643e945692c3 req-1e2b1f71-b445-4c5d-a0c8-935f00452ce3 service nova] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Updated VIF entry in instance network info cache for port 81ffb8f2-c59b-4b0e-894a-d8633440d6eb. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 997.347197] env[61728]: DEBUG nova.network.neutron [req-d0f7e9b6-1828-4cdf-b5ef-643e945692c3 req-1e2b1f71-b445-4c5d-a0c8-935f00452ce3 service nova] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Updating instance_info_cache with network_info: [{"id": "81ffb8f2-c59b-4b0e-894a-d8633440d6eb", "address": "fa:16:3e:cb:69:dc", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81ffb8f2-c5", "ovs_interfaceid": "81ffb8f2-c59b-4b0e-894a-d8633440d6eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.397356] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52548d9c-0f34-d719-2d6a-e131191fdb55, 'name': SearchDatastore_Task, 'duration_secs': 0.014636} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.397689] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.397926] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 997.398246] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.398428] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.398686] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 997.398959] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bbe59399-9199-4027-a30f-cd33d3a9c75b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.420709] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 997.420833] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 997.421588] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98c3c9a9-1d0a-4e54-ba9f-1af22b5f2136 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.428195] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 997.428195] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524cf515-ad3a-c484-0eaf-b34c0cb0218c" [ 997.428195] env[61728]: _type = "Task" [ 997.428195] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.438644] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524cf515-ad3a-c484-0eaf-b34c0cb0218c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.439060] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6b6d29bb-8df3-404b-a953-d9f4006a21da tempest-ImagesTestJSON-1033899734 tempest-ImagesTestJSON-1033899734-project-member] Lock "0af88a53-bbe7-443e-a7dc-734e483bb730" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.354s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.478568] env[61728]: DEBUG oslo_vmware.api [None req-12fcc626-3d68-414c-9d1a-78999e439a45 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464652, 'name': SuspendVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.724992] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "interface-a70d1c37-e792-4168-b7e1-9418b8cb7818-d0b341ea-4a0b-46ff-89f2-f96d212d04d9" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.725618] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-a70d1c37-e792-4168-b7e1-9418b8cb7818-d0b341ea-4a0b-46ff-89f2-f96d212d04d9" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.725756] env[61728]: DEBUG nova.objects.instance [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'flavor' on Instance uuid a70d1c37-e792-4168-b7e1-9418b8cb7818 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.850199] env[61728]: DEBUG oslo_concurrency.lockutils [req-d0f7e9b6-1828-4cdf-b5ef-643e945692c3 req-1e2b1f71-b445-4c5d-a0c8-935f00452ce3 service nova] Releasing lock "refresh_cache-b4241d6f-1d70-4f61-a322-caff7c4e6f64" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.942684] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524cf515-ad3a-c484-0eaf-b34c0cb0218c, 'name': SearchDatastore_Task, 'duration_secs': 0.019185} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.948504] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93d790b1-4836-48a9-a70f-5e43d2d02e9f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.957839] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 997.957839] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a99d70-485d-e138-e95a-d472b3035707" [ 997.957839] env[61728]: _type = "Task" [ 997.957839] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.966331] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a99d70-485d-e138-e95a-d472b3035707, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.978913] env[61728]: DEBUG oslo_vmware.api [None req-12fcc626-3d68-414c-9d1a-78999e439a45 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464652, 'name': SuspendVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.980965] env[61728]: DEBUG nova.compute.manager [req-b2267822-6382-4894-a01a-35e117211af7 req-51b0d504-37a4-48d3-a658-7c00974abf5f service nova] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Received event network-vif-plugged-eb281159-3780-4ea6-b463-cd8d839be907 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 997.981198] env[61728]: DEBUG oslo_concurrency.lockutils [req-b2267822-6382-4894-a01a-35e117211af7 req-51b0d504-37a4-48d3-a658-7c00974abf5f service nova] Acquiring lock "267880ff-0fac-4763-82be-ef8587ee7628-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.981407] env[61728]: DEBUG oslo_concurrency.lockutils [req-b2267822-6382-4894-a01a-35e117211af7 req-51b0d504-37a4-48d3-a658-7c00974abf5f service nova] Lock "267880ff-0fac-4763-82be-ef8587ee7628-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.981578] env[61728]: DEBUG oslo_concurrency.lockutils [req-b2267822-6382-4894-a01a-35e117211af7 req-51b0d504-37a4-48d3-a658-7c00974abf5f service nova] Lock "267880ff-0fac-4763-82be-ef8587ee7628-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.981776] env[61728]: DEBUG nova.compute.manager [req-b2267822-6382-4894-a01a-35e117211af7 req-51b0d504-37a4-48d3-a658-7c00974abf5f service nova] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] No waiting events found dispatching network-vif-plugged-eb281159-3780-4ea6-b463-cd8d839be907 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 997.981963] env[61728]: WARNING nova.compute.manager [req-b2267822-6382-4894-a01a-35e117211af7 req-51b0d504-37a4-48d3-a658-7c00974abf5f service nova] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Received unexpected event network-vif-plugged-eb281159-3780-4ea6-b463-cd8d839be907 for instance with vm_state building and task_state spawning. [ 998.181434] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ca54ba-5f24-42ac-a6e4-b21dc6709394 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.189308] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8558d00d-6e8e-4665-af49-21386e0d16da {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.222820] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee661d45-b6de-4179-b93e-4647d5ece481 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.233683] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21313b1d-7217-40ec-b1d3-4b2eb5eab927 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.251396] env[61728]: DEBUG nova.compute.provider_tree [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.268605] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 998.268837] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 998.269034] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleting the datastore file [datastore1] 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.269559] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3fbd718b-8726-4298-b9b0-d58dd9ed44aa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.276375] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 998.276375] env[61728]: value = "task-464655" [ 998.276375] env[61728]: _type = "Task" [ 998.276375] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.284922] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464655, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.348230] env[61728]: DEBUG nova.objects.instance [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'pci_requests' on Instance uuid a70d1c37-e792-4168-b7e1-9418b8cb7818 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.468383] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a99d70-485d-e138-e95a-d472b3035707, 'name': SearchDatastore_Task, 'duration_secs': 0.013514} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.468534] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.468798] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b4241d6f-1d70-4f61-a322-caff7c4e6f64/b4241d6f-1d70-4f61-a322-caff7c4e6f64.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 998.469078] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40e7d38f-55b4-45b2-ad33-062d12f073e6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.480523] env[61728]: DEBUG oslo_vmware.api [None req-12fcc626-3d68-414c-9d1a-78999e439a45 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464652, 'name': SuspendVM_Task, 'duration_secs': 1.400404} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.481820] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-12fcc626-3d68-414c-9d1a-78999e439a45 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Suspended the VM {{(pid=61728) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 998.482396] env[61728]: DEBUG nova.compute.manager [None req-12fcc626-3d68-414c-9d1a-78999e439a45 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.482396] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 998.482396] env[61728]: value = "task-464656" [ 998.482396] env[61728]: _type = "Task" [ 998.482396] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.483086] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fecc36b-9be2-4ce9-bdfc-21478161d6d1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.489938] env[61728]: DEBUG nova.network.neutron [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Successfully updated port: eb281159-3780-4ea6-b463-cd8d839be907 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 998.497613] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464656, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.755290] env[61728]: DEBUG nova.scheduler.client.report [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.765695] env[61728]: DEBUG nova.compute.manager [req-bd5826c9-459b-4a39-9d22-f6a9ec3a2799 req-7494fbd1-2c76-462d-8d66-40b5cb1c19b8 service nova] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Received event network-changed-eb281159-3780-4ea6-b463-cd8d839be907 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 998.765954] env[61728]: DEBUG nova.compute.manager [req-bd5826c9-459b-4a39-9d22-f6a9ec3a2799 req-7494fbd1-2c76-462d-8d66-40b5cb1c19b8 service nova] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Refreshing instance network info cache due to event network-changed-eb281159-3780-4ea6-b463-cd8d839be907. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 998.766332] env[61728]: DEBUG oslo_concurrency.lockutils [req-bd5826c9-459b-4a39-9d22-f6a9ec3a2799 req-7494fbd1-2c76-462d-8d66-40b5cb1c19b8 service nova] Acquiring lock "refresh_cache-267880ff-0fac-4763-82be-ef8587ee7628" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.766533] env[61728]: DEBUG oslo_concurrency.lockutils [req-bd5826c9-459b-4a39-9d22-f6a9ec3a2799 req-7494fbd1-2c76-462d-8d66-40b5cb1c19b8 service nova] Acquired lock "refresh_cache-267880ff-0fac-4763-82be-ef8587ee7628" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.766746] env[61728]: DEBUG nova.network.neutron [req-bd5826c9-459b-4a39-9d22-f6a9ec3a2799 req-7494fbd1-2c76-462d-8d66-40b5cb1c19b8 service nova] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Refreshing network info cache for port eb281159-3780-4ea6-b463-cd8d839be907 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 998.790379] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464655, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.202259} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.790677] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 998.790869] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 998.791134] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 998.851117] env[61728]: DEBUG nova.objects.base [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 998.851367] env[61728]: DEBUG nova.network.neutron [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 998.963935] env[61728]: DEBUG nova.policy [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e51efcf23db4d5496edbc3b7104b4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32bb35e1dfce40e48be08bb568d3f2b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 999.000722] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "refresh_cache-267880ff-0fac-4763-82be-ef8587ee7628" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.000722] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464656, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.449274} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.000722] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] b4241d6f-1d70-4f61-a322-caff7c4e6f64/b4241d6f-1d70-4f61-a322-caff7c4e6f64.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 999.000722] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 999.001215] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9b155e4c-65f8-460a-8f85-abc28d648a4a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.010646] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 999.010646] env[61728]: value = "task-464657" [ 999.010646] env[61728]: _type = "Task" [ 999.010646] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.020517] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464657, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.263024] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.351s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.263024] env[61728]: DEBUG nova.compute.manager [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 999.264377] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.163s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.264741] env[61728]: DEBUG nova.objects.instance [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lazy-loading 'resources' on Instance uuid 0ff847f9-97d0-40d9-a458-038fe36b5c62 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.331513] env[61728]: DEBUG nova.network.neutron [req-bd5826c9-459b-4a39-9d22-f6a9ec3a2799 req-7494fbd1-2c76-462d-8d66-40b5cb1c19b8 service nova] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 999.377279] env[61728]: DEBUG nova.network.neutron [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Port 83657468-5cac-4258-84f4-b105395d9cfa binding to destination host cpu-1 is already ACTIVE {{(pid=61728) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 999.377554] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.377715] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.377873] env[61728]: DEBUG nova.network.neutron [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 999.480937] env[61728]: DEBUG nova.network.neutron [req-bd5826c9-459b-4a39-9d22-f6a9ec3a2799 req-7494fbd1-2c76-462d-8d66-40b5cb1c19b8 service nova] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.525251] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464657, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069183} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.525251] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 999.525445] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c811696a-9d35-4433-bdf5-c35ffd548c05 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.549239] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] b4241d6f-1d70-4f61-a322-caff7c4e6f64/b4241d6f-1d70-4f61-a322-caff7c4e6f64.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.549902] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e171e7ec-c535-43d9-a63b-ff2a08d63417 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.571665] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 999.571665] env[61728]: value = "task-464658" [ 999.571665] env[61728]: _type = "Task" [ 999.571665] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.581842] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464658, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.768210] env[61728]: DEBUG nova.compute.utils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 999.772634] env[61728]: DEBUG nova.compute.manager [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 999.772843] env[61728]: DEBUG nova.network.neutron [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 999.782771] env[61728]: DEBUG oslo_concurrency.lockutils [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "172129ff-136e-4855-8659-3f1ac88dcbd2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.787285] env[61728]: DEBUG oslo_concurrency.lockutils [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "172129ff-136e-4855-8659-3f1ac88dcbd2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.787285] env[61728]: DEBUG oslo_concurrency.lockutils [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "172129ff-136e-4855-8659-3f1ac88dcbd2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.787285] env[61728]: DEBUG oslo_concurrency.lockutils [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "172129ff-136e-4855-8659-3f1ac88dcbd2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.787285] env[61728]: DEBUG oslo_concurrency.lockutils [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "172129ff-136e-4855-8659-3f1ac88dcbd2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.789110] env[61728]: INFO nova.compute.manager [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Terminating instance [ 999.793467] env[61728]: DEBUG nova.compute.manager [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 999.793467] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 999.793467] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58cfec61-ec5f-4d11-9349-2d882cc9ea44 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.801027] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 999.801228] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1352f42-2774-410a-b0f5-0641a59d6621 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.837561] env[61728]: DEBUG nova.policy [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '38f64fd31712472a841f82c620d9340b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f5329f59d8bf43b2a855c4259387fccd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 999.842663] env[61728]: DEBUG nova.virt.hardware [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 999.842950] env[61728]: DEBUG nova.virt.hardware [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 999.843192] env[61728]: DEBUG nova.virt.hardware [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 999.843422] env[61728]: DEBUG nova.virt.hardware [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 999.843614] env[61728]: DEBUG nova.virt.hardware [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 999.843778] env[61728]: DEBUG nova.virt.hardware [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 999.844047] env[61728]: DEBUG nova.virt.hardware [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 999.844282] env[61728]: DEBUG nova.virt.hardware [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 999.844396] env[61728]: DEBUG nova.virt.hardware [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 999.844765] env[61728]: DEBUG nova.virt.hardware [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 999.845017] env[61728]: DEBUG nova.virt.hardware [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 999.845954] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8af183a-2597-4614-a19c-bae66172212a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.854969] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9959c01-62d0-4675-b9cc-a9369e0ca87a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.869071] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:17:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9732690c-bdcf-4e6f-9a32-42c196333eb8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '18a00538-a062-4136-b2d4-2b3b3f783502', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 999.876735] env[61728]: DEBUG oslo.service.loopingcall [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.879309] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 999.879736] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7d37b0b3-a1e5-4f75-b50a-af5b7a2e363f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.904115] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 999.904115] env[61728]: value = "task-464660" [ 999.904115] env[61728]: _type = "Task" [ 999.904115] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.913615] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464660, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.932129] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 999.932367] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 999.932568] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleting the datastore file [datastore1] 172129ff-136e-4855-8659-3f1ac88dcbd2 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 999.932865] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bff62c2d-2244-4670-a78d-b079b1d70e10 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.941574] env[61728]: DEBUG oslo_vmware.api [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 999.941574] env[61728]: value = "task-464661" [ 999.941574] env[61728]: _type = "Task" [ 999.941574] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.960037] env[61728]: DEBUG oslo_vmware.api [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464661, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.984263] env[61728]: DEBUG oslo_concurrency.lockutils [req-bd5826c9-459b-4a39-9d22-f6a9ec3a2799 req-7494fbd1-2c76-462d-8d66-40b5cb1c19b8 service nova] Releasing lock "refresh_cache-267880ff-0fac-4763-82be-ef8587ee7628" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.984263] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired lock "refresh_cache-267880ff-0fac-4763-82be-ef8587ee7628" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.984263] env[61728]: DEBUG nova.network.neutron [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1000.083679] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464658, 'name': ReconfigVM_Task, 'duration_secs': 0.285553} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.089725] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Reconfigured VM instance instance-00000063 to attach disk [datastore1] b4241d6f-1d70-4f61-a322-caff7c4e6f64/b4241d6f-1d70-4f61-a322-caff7c4e6f64.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.090652] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd9948ce-7a77-47e9-b7bf-0a4b8818e2db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.098863] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1000.098863] env[61728]: value = "task-464662" [ 1000.098863] env[61728]: _type = "Task" [ 1000.098863] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.113862] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464662, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.135070] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efccab5f-6ec3-47e9-9562-c4ab877281d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.144976] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e2407a-b191-43c4-bd62-4ac4c64c4a29 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.180450] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e9c4ec-fed2-42a3-a41c-d3e124df4ae1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.189103] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2324fa4d-77c1-443b-a8cc-5024e5dfada9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.205387] env[61728]: DEBUG nova.compute.provider_tree [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1000.273689] env[61728]: DEBUG nova.compute.manager [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1000.306632] env[61728]: DEBUG nova.network.neutron [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance_info_cache with network_info: [{"id": "83657468-5cac-4258-84f4-b105395d9cfa", "address": "fa:16:3e:55:1a:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83657468-5c", "ovs_interfaceid": "83657468-5cac-4258-84f4-b105395d9cfa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.316859] env[61728]: DEBUG nova.network.neutron [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Successfully created port: f99497f4-e49f-4d6b-9bc8-75efe36beb4b {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1000.415970] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464660, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.452050] env[61728]: DEBUG oslo_vmware.api [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464661, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.385195} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.452396] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.452639] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1000.452926] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1000.453131] env[61728]: INFO nova.compute.manager [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Took 0.66 seconds to destroy the instance on the hypervisor. [ 1000.453380] env[61728]: DEBUG oslo.service.loopingcall [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.453577] env[61728]: DEBUG nova.compute.manager [-] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1000.453666] env[61728]: DEBUG nova.network.neutron [-] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1000.569970] env[61728]: DEBUG nova.network.neutron [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1000.613416] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464662, 'name': Rename_Task, 'duration_secs': 0.149032} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.617019] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1000.617019] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f44fd3d0-fb26-419d-b17e-9374dfe4b2ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.621664] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1000.621664] env[61728]: value = "task-464663" [ 1000.621664] env[61728]: _type = "Task" [ 1000.621664] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.637285] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464663, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.725505] env[61728]: DEBUG nova.network.neutron [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Successfully updated port: d0b341ea-4a0b-46ff-89f2-f96d212d04d9 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1000.763553] env[61728]: DEBUG nova.scheduler.client.report [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 151 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1000.763836] env[61728]: DEBUG nova.compute.provider_tree [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 151 to 152 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1000.764037] env[61728]: DEBUG nova.compute.provider_tree [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1000.811855] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.915171] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464660, 'name': CreateVM_Task, 'duration_secs': 0.565807} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.915417] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1000.916158] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.916391] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.916693] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1000.916951] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b9450fd-c6ba-4742-9c40-2a0a85206361 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.922817] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1000.922817] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aaaad3-29e9-d7b0-0d16-4f5612f4c544" [ 1000.922817] env[61728]: _type = "Task" [ 1000.922817] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.931432] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aaaad3-29e9-d7b0-0d16-4f5612f4c544, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.030986] env[61728]: DEBUG nova.compute.manager [req-ff1226d9-05f2-4e87-8054-86228825870b req-15cf790e-95fb-47fb-975d-35097ebf8f41 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Received event network-vif-plugged-d0b341ea-4a0b-46ff-89f2-f96d212d04d9 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1001.031253] env[61728]: DEBUG oslo_concurrency.lockutils [req-ff1226d9-05f2-4e87-8054-86228825870b req-15cf790e-95fb-47fb-975d-35097ebf8f41 service nova] Acquiring lock "a70d1c37-e792-4168-b7e1-9418b8cb7818-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.031462] env[61728]: DEBUG oslo_concurrency.lockutils [req-ff1226d9-05f2-4e87-8054-86228825870b req-15cf790e-95fb-47fb-975d-35097ebf8f41 service nova] Lock "a70d1c37-e792-4168-b7e1-9418b8cb7818-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.031785] env[61728]: DEBUG oslo_concurrency.lockutils [req-ff1226d9-05f2-4e87-8054-86228825870b req-15cf790e-95fb-47fb-975d-35097ebf8f41 service nova] Lock "a70d1c37-e792-4168-b7e1-9418b8cb7818-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.032118] env[61728]: DEBUG nova.compute.manager [req-ff1226d9-05f2-4e87-8054-86228825870b req-15cf790e-95fb-47fb-975d-35097ebf8f41 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] No waiting events found dispatching network-vif-plugged-d0b341ea-4a0b-46ff-89f2-f96d212d04d9 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1001.032331] env[61728]: WARNING nova.compute.manager [req-ff1226d9-05f2-4e87-8054-86228825870b req-15cf790e-95fb-47fb-975d-35097ebf8f41 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Received unexpected event network-vif-plugged-d0b341ea-4a0b-46ff-89f2-f96d212d04d9 for instance with vm_state active and task_state None. [ 1001.032791] env[61728]: DEBUG nova.compute.manager [req-ff1226d9-05f2-4e87-8054-86228825870b req-15cf790e-95fb-47fb-975d-35097ebf8f41 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Received event network-changed-d0b341ea-4a0b-46ff-89f2-f96d212d04d9 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1001.033132] env[61728]: DEBUG nova.compute.manager [req-ff1226d9-05f2-4e87-8054-86228825870b req-15cf790e-95fb-47fb-975d-35097ebf8f41 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Refreshing instance network info cache due to event network-changed-d0b341ea-4a0b-46ff-89f2-f96d212d04d9. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1001.033525] env[61728]: DEBUG oslo_concurrency.lockutils [req-ff1226d9-05f2-4e87-8054-86228825870b req-15cf790e-95fb-47fb-975d-35097ebf8f41 service nova] Acquiring lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.033711] env[61728]: DEBUG oslo_concurrency.lockutils [req-ff1226d9-05f2-4e87-8054-86228825870b req-15cf790e-95fb-47fb-975d-35097ebf8f41 service nova] Acquired lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.033883] env[61728]: DEBUG nova.network.neutron [req-ff1226d9-05f2-4e87-8054-86228825870b req-15cf790e-95fb-47fb-975d-35097ebf8f41 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Refreshing network info cache for port d0b341ea-4a0b-46ff-89f2-f96d212d04d9 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1001.084940] env[61728]: DEBUG nova.network.neutron [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Updating instance_info_cache with network_info: [{"id": "eb281159-3780-4ea6-b463-cd8d839be907", "address": "fa:16:3e:ff:b3:0b", "network": {"id": "22ad4b68-2895-4651-803f-44b8031a6292", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-489227259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "276942c8cf8a42729a541de096f69a0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb281159-37", "ovs_interfaceid": "eb281159-3780-4ea6-b463-cd8d839be907", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.142741] env[61728]: DEBUG oslo_vmware.api [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464663, 'name': PowerOnVM_Task, 'duration_secs': 0.507146} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.143460] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1001.143709] env[61728]: INFO nova.compute.manager [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Took 6.64 seconds to spawn the instance on the hypervisor. [ 1001.143986] env[61728]: DEBUG nova.compute.manager [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.144931] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6f251f-915d-4eb5-9190-be4dd374690f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.230589] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.269168] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.005s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.271971] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.164s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.272211] env[61728]: DEBUG nova.objects.instance [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lazy-loading 'resources' on Instance uuid db547fea-4d44-440e-ac3e-29ae9b2c547b {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.285604] env[61728]: DEBUG nova.compute.manager [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1001.293189] env[61728]: INFO nova.scheduler.client.report [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Deleted allocations for instance 0ff847f9-97d0-40d9-a458-038fe36b5c62 [ 1001.313108] env[61728]: DEBUG nova.virt.hardware [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1001.313558] env[61728]: DEBUG nova.virt.hardware [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1001.313777] env[61728]: DEBUG nova.virt.hardware [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1001.314086] env[61728]: DEBUG nova.virt.hardware [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1001.314212] env[61728]: DEBUG nova.virt.hardware [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1001.314408] env[61728]: DEBUG nova.virt.hardware [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1001.314676] env[61728]: DEBUG nova.virt.hardware [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1001.314894] env[61728]: DEBUG nova.virt.hardware [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1001.315727] env[61728]: DEBUG nova.virt.hardware [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1001.315964] env[61728]: DEBUG nova.virt.hardware [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1001.316211] env[61728]: DEBUG nova.virt.hardware [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1001.317472] env[61728]: DEBUG nova.compute.manager [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61728) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1001.319875] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367cca15-7d33-47d7-bc13-024b632117a3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.335602] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4caac9ad-a807-4dad-ac72-fddd927f9681 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.435283] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aaaad3-29e9-d7b0-0d16-4f5612f4c544, 'name': SearchDatastore_Task, 'duration_secs': 0.011649} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.435616] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.435856] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1001.436296] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.436296] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.436537] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1001.436813] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b1258165-4eb0-499c-ab5a-c463cac61823 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.450841] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1001.451184] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1001.451871] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7560954-9a62-4600-81f2-6eb50fa550e6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.457985] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1001.457985] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528d1cc3-9244-4e3a-b74f-79415f049d7e" [ 1001.457985] env[61728]: _type = "Task" [ 1001.457985] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.465984] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528d1cc3-9244-4e3a-b74f-79415f049d7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.525601] env[61728]: DEBUG nova.compute.manager [req-a83413b2-3d99-4c6f-ba0e-23a96a93e760 req-359eac8c-ee63-4944-80f7-1574611d2aca service nova] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Received event network-vif-deleted-03e4158f-eab0-4cd7-9d4e-b8ee77e099c6 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1001.525741] env[61728]: INFO nova.compute.manager [req-a83413b2-3d99-4c6f-ba0e-23a96a93e760 req-359eac8c-ee63-4944-80f7-1574611d2aca service nova] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Neutron deleted interface 03e4158f-eab0-4cd7-9d4e-b8ee77e099c6; detaching it from the instance and deleting it from the info cache [ 1001.526025] env[61728]: DEBUG nova.network.neutron [req-a83413b2-3d99-4c6f-ba0e-23a96a93e760 req-359eac8c-ee63-4944-80f7-1574611d2aca service nova] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.588173] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Releasing lock "refresh_cache-267880ff-0fac-4763-82be-ef8587ee7628" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.588489] env[61728]: DEBUG nova.compute.manager [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Instance network_info: |[{"id": "eb281159-3780-4ea6-b463-cd8d839be907", "address": "fa:16:3e:ff:b3:0b", "network": {"id": "22ad4b68-2895-4651-803f-44b8031a6292", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-489227259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "276942c8cf8a42729a541de096f69a0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb281159-37", "ovs_interfaceid": "eb281159-3780-4ea6-b463-cd8d839be907", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1001.588896] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:b3:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0248a27a-1d7f-4195-987b-06bfc8467347', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb281159-3780-4ea6-b463-cd8d839be907', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1001.597053] env[61728]: DEBUG oslo.service.loopingcall [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.597559] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1001.597794] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6edeb7cd-3b30-4e6a-ae12-89322887887d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.619615] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1001.619615] env[61728]: value = "task-464664" [ 1001.619615] env[61728]: _type = "Task" [ 1001.619615] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.627442] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464664, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.671566] env[61728]: INFO nova.compute.manager [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Took 26.05 seconds to build instance. [ 1001.691434] env[61728]: DEBUG nova.network.neutron [-] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.803687] env[61728]: DEBUG nova.network.neutron [req-ff1226d9-05f2-4e87-8054-86228825870b req-15cf790e-95fb-47fb-975d-35097ebf8f41 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Added VIF to instance network info cache for port d0b341ea-4a0b-46ff-89f2-f96d212d04d9. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3489}} [ 1001.804120] env[61728]: DEBUG nova.network.neutron [req-ff1226d9-05f2-4e87-8054-86228825870b req-15cf790e-95fb-47fb-975d-35097ebf8f41 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updating instance_info_cache with network_info: [{"id": "c8407eac-4337-4ae3-a071-08f48bf87248", "address": "fa:16:3e:2d:a6:d6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8407eac-43", "ovs_interfaceid": "c8407eac-4337-4ae3-a071-08f48bf87248", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d0b341ea-4a0b-46ff-89f2-f96d212d04d9", "address": "fa:16:3e:0a:de:a7", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0b341ea-4a", "ovs_interfaceid": "d0b341ea-4a0b-46ff-89f2-f96d212d04d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.807060] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9b486039-d83d-427d-824a-57556e17135a tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "0ff847f9-97d0-40d9-a458-038fe36b5c62" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.276s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.970849] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528d1cc3-9244-4e3a-b74f-79415f049d7e, 'name': SearchDatastore_Task, 'duration_secs': 0.031258} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.975067] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0b477fe-012c-4bce-b35e-691fb244ae0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.982126] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1001.982126] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5280331a-4979-5ccd-6ee1-ef1922377e8f" [ 1001.982126] env[61728]: _type = "Task" [ 1001.982126] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.999862] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5280331a-4979-5ccd-6ee1-ef1922377e8f, 'name': SearchDatastore_Task, 'duration_secs': 0.010335} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.999862] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.000134] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f/0edd1f56-3e8e-4868-9c8e-a8caaf026e7f.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1002.000806] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1a9a341e-0ffe-40dc-9832-4712e217f6d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.010434] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1002.010434] env[61728]: value = "task-464665" [ 1002.010434] env[61728]: _type = "Task" [ 1002.010434] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.021191] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464665, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.029122] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-20227bf0-bab2-474e-8179-14dcb719f090 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.044850] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cea172-6097-47c2-910f-beffdf9b7eec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.074719] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec5193be-2a7f-4954-b523-f39735fba5a4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.089440] env[61728]: DEBUG nova.compute.manager [req-a83413b2-3d99-4c6f-ba0e-23a96a93e760 req-359eac8c-ee63-4944-80f7-1574611d2aca service nova] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Detach interface failed, port_id=03e4158f-eab0-4cd7-9d4e-b8ee77e099c6, reason: Instance 172129ff-136e-4855-8659-3f1ac88dcbd2 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1002.095381] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5255f6d-6a2f-4c64-8d6e-931f1730abe1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.133726] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c8bfc7-5103-4eb5-abb1-b4bb17bc5d31 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.142628] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464664, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.146111] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42ee4dc-caa0-42ab-9dd9-9094e9ad6d25 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.164565] env[61728]: DEBUG nova.compute.provider_tree [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.173315] env[61728]: DEBUG oslo_concurrency.lockutils [None req-449c1cfc-702a-4da6-862f-ce00ea74a4fb tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "b4241d6f-1d70-4f61-a322-caff7c4e6f64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.569s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.194779] env[61728]: INFO nova.compute.manager [-] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Took 1.74 seconds to deallocate network for instance. [ 1002.289595] env[61728]: DEBUG nova.network.neutron [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Successfully updated port: f99497f4-e49f-4d6b-9bc8-75efe36beb4b {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1002.307406] env[61728]: DEBUG oslo_concurrency.lockutils [req-ff1226d9-05f2-4e87-8054-86228825870b req-15cf790e-95fb-47fb-975d-35097ebf8f41 service nova] Releasing lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.307867] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.308169] env[61728]: DEBUG nova.network.neutron [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1002.448355] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.526256] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464665, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.641764] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464664, 'name': CreateVM_Task, 'duration_secs': 0.564246} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.641958] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1002.642704] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.642920] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.643301] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1002.643584] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3dfbbf5d-0c7c-47e5-9ded-9d3294d2e500 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.649816] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1002.649816] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526a2b03-cf30-88db-e9da-0695fd2c4d45" [ 1002.649816] env[61728]: _type = "Task" [ 1002.649816] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.658691] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526a2b03-cf30-88db-e9da-0695fd2c4d45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.669918] env[61728]: DEBUG nova.scheduler.client.report [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1002.701525] env[61728]: DEBUG oslo_concurrency.lockutils [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.718112] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "bb7b38d5-53c7-4c05-96f8-9121171d3c46" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.718519] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "bb7b38d5-53c7-4c05-96f8-9121171d3c46" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.792488] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "refresh_cache-378ef752-6d47-4861-baf1-064b2a180dcc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.792602] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "refresh_cache-378ef752-6d47-4861-baf1-064b2a180dcc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.792778] env[61728]: DEBUG nova.network.neutron [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1002.849067] env[61728]: WARNING nova.network.neutron [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] 20ba9b61-3be9-4ad2-a1d0-88c810873bcb already exists in list: networks containing: ['20ba9b61-3be9-4ad2-a1d0-88c810873bcb']. ignoring it [ 1002.849277] env[61728]: WARNING nova.network.neutron [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] 20ba9b61-3be9-4ad2-a1d0-88c810873bcb already exists in list: networks containing: ['20ba9b61-3be9-4ad2-a1d0-88c810873bcb']. ignoring it [ 1002.849452] env[61728]: WARNING nova.network.neutron [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] d0b341ea-4a0b-46ff-89f2-f96d212d04d9 already exists in list: port_ids containing: ['d0b341ea-4a0b-46ff-89f2-f96d212d04d9']. ignoring it [ 1003.024032] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464665, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.626573} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.024528] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f/0edd1f56-3e8e-4868-9c8e-a8caaf026e7f.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1003.026365] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1003.026365] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fb623cce-214f-49cf-96be-3945f2976ccf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.033297] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1003.033297] env[61728]: value = "task-464666" [ 1003.033297] env[61728]: _type = "Task" [ 1003.033297] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.041571] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464666, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.147878] env[61728]: DEBUG nova.network.neutron [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updating instance_info_cache with network_info: [{"id": "c8407eac-4337-4ae3-a071-08f48bf87248", "address": "fa:16:3e:2d:a6:d6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8407eac-43", "ovs_interfaceid": "c8407eac-4337-4ae3-a071-08f48bf87248", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d0b341ea-4a0b-46ff-89f2-f96d212d04d9", "address": "fa:16:3e:0a:de:a7", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0b341ea-4a", "ovs_interfaceid": "d0b341ea-4a0b-46ff-89f2-f96d212d04d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.161138] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526a2b03-cf30-88db-e9da-0695fd2c4d45, 'name': SearchDatastore_Task, 'duration_secs': 0.030581} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.162161] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.162405] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1003.162700] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.162925] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.163245] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1003.163785] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f9e5f03d-9bde-41f7-8fbc-f2adf68353f7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.175250] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.904s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.177966] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.066s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.178061] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.179797] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.991s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.180052] env[61728]: DEBUG nova.objects.instance [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Lazy-loading 'resources' on Instance uuid 7e8cde9c-890c-4f50-9ccb-46aa71220395 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.181125] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1003.181308] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1003.184455] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72429ccd-7613-40f5-90f9-e0bec170f185 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.190874] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1003.190874] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524b4a99-9f14-8687-ab16-c19d3d742809" [ 1003.190874] env[61728]: _type = "Task" [ 1003.190874] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.201318] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524b4a99-9f14-8687-ab16-c19d3d742809, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.207352] env[61728]: INFO nova.scheduler.client.report [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Deleted allocations for instance db547fea-4d44-440e-ac3e-29ae9b2c547b [ 1003.209011] env[61728]: INFO nova.scheduler.client.report [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleted allocations for instance 3415687e-9b8d-42f8-9b18-93951b8ec7da [ 1003.220200] env[61728]: DEBUG nova.compute.manager [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1003.329983] env[61728]: DEBUG nova.network.neutron [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1003.545846] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464666, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.20589} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.546232] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1003.546959] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-786bc990-d6f0-430a-b8a9-73f0eae6f243 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.575339] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f/0edd1f56-3e8e-4868-9c8e-a8caaf026e7f.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1003.576464] env[61728]: DEBUG nova.network.neutron [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Updating instance_info_cache with network_info: [{"id": "f99497f4-e49f-4d6b-9bc8-75efe36beb4b", "address": "fa:16:3e:87:0a:68", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf99497f4-e4", "ovs_interfaceid": "f99497f4-e49f-4d6b-9bc8-75efe36beb4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.577671] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62bf7fd4-a9bb-4192-8a36-ef3fd9f7f4ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.599714] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1003.599714] env[61728]: value = "task-464667" [ 1003.599714] env[61728]: _type = "Task" [ 1003.599714] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.610965] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464667, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.649313] env[61728]: DEBUG nova.compute.manager [req-440250c4-5fbd-4b7b-96bc-e8a745a8db6b req-3b9bc160-a625-4478-adf2-9b7c281ec6da service nova] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Received event network-vif-plugged-f99497f4-e49f-4d6b-9bc8-75efe36beb4b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1003.649538] env[61728]: DEBUG oslo_concurrency.lockutils [req-440250c4-5fbd-4b7b-96bc-e8a745a8db6b req-3b9bc160-a625-4478-adf2-9b7c281ec6da service nova] Acquiring lock "378ef752-6d47-4861-baf1-064b2a180dcc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.649799] env[61728]: DEBUG oslo_concurrency.lockutils [req-440250c4-5fbd-4b7b-96bc-e8a745a8db6b req-3b9bc160-a625-4478-adf2-9b7c281ec6da service nova] Lock "378ef752-6d47-4861-baf1-064b2a180dcc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.649979] env[61728]: DEBUG oslo_concurrency.lockutils [req-440250c4-5fbd-4b7b-96bc-e8a745a8db6b req-3b9bc160-a625-4478-adf2-9b7c281ec6da service nova] Lock "378ef752-6d47-4861-baf1-064b2a180dcc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.650427] env[61728]: DEBUG nova.compute.manager [req-440250c4-5fbd-4b7b-96bc-e8a745a8db6b req-3b9bc160-a625-4478-adf2-9b7c281ec6da service nova] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] No waiting events found dispatching network-vif-plugged-f99497f4-e49f-4d6b-9bc8-75efe36beb4b {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1003.650613] env[61728]: WARNING nova.compute.manager [req-440250c4-5fbd-4b7b-96bc-e8a745a8db6b req-3b9bc160-a625-4478-adf2-9b7c281ec6da service nova] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Received unexpected event network-vif-plugged-f99497f4-e49f-4d6b-9bc8-75efe36beb4b for instance with vm_state building and task_state spawning. [ 1003.650782] env[61728]: DEBUG nova.compute.manager [req-440250c4-5fbd-4b7b-96bc-e8a745a8db6b req-3b9bc160-a625-4478-adf2-9b7c281ec6da service nova] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Received event network-changed-f99497f4-e49f-4d6b-9bc8-75efe36beb4b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1003.650944] env[61728]: DEBUG nova.compute.manager [req-440250c4-5fbd-4b7b-96bc-e8a745a8db6b req-3b9bc160-a625-4478-adf2-9b7c281ec6da service nova] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Refreshing instance network info cache due to event network-changed-f99497f4-e49f-4d6b-9bc8-75efe36beb4b. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1003.651825] env[61728]: DEBUG oslo_concurrency.lockutils [req-440250c4-5fbd-4b7b-96bc-e8a745a8db6b req-3b9bc160-a625-4478-adf2-9b7c281ec6da service nova] Acquiring lock "refresh_cache-378ef752-6d47-4861-baf1-064b2a180dcc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.652080] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.652700] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.652888] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.653841] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d688477a-8ac5-4aa9-926f-aa3a335f533c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.673520] env[61728]: DEBUG nova.virt.hardware [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1003.673803] env[61728]: DEBUG nova.virt.hardware [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1003.673994] env[61728]: DEBUG nova.virt.hardware [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1003.674263] env[61728]: DEBUG nova.virt.hardware [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1003.674431] env[61728]: DEBUG nova.virt.hardware [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1003.674633] env[61728]: DEBUG nova.virt.hardware [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1003.674874] env[61728]: DEBUG nova.virt.hardware [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1003.674993] env[61728]: DEBUG nova.virt.hardware [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1003.675392] env[61728]: DEBUG nova.virt.hardware [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1003.675478] env[61728]: DEBUG nova.virt.hardware [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1003.676245] env[61728]: DEBUG nova.virt.hardware [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1003.682419] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Reconfiguring VM to attach interface {{(pid=61728) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1003.683175] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-afaaf933-0586-43b3-ad63-8bf4e873e77d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.710137] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524b4a99-9f14-8687-ab16-c19d3d742809, 'name': SearchDatastore_Task, 'duration_secs': 0.01442} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.715023] env[61728]: DEBUG oslo_vmware.api [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 1003.715023] env[61728]: value = "task-464668" [ 1003.715023] env[61728]: _type = "Task" [ 1003.715023] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.715023] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aea773bf-01b8-4a31-baa1-71a86cedcefe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.732292] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dff5cc8e-cf54-4bc9-857c-50434a421687 tempest-MultipleCreateTestJSON-1855804121 tempest-MultipleCreateTestJSON-1855804121-project-member] Lock "db547fea-4d44-440e-ac3e-29ae9b2c547b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.406s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.735761] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1003.735761] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528fa4b3-de42-f80c-35b1-749af6cd5e39" [ 1003.735761] env[61728]: _type = "Task" [ 1003.735761] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.738728] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6a7c2297-8392-4e50-8172-facb4e5dcfdd tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "3415687e-9b8d-42f8-9b18-93951b8ec7da" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.924s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.748877] env[61728]: DEBUG oslo_vmware.api [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464668, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.754432] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528fa4b3-de42-f80c-35b1-749af6cd5e39, 'name': SearchDatastore_Task, 'duration_secs': 0.017896} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.754694] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.755853] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 267880ff-0fac-4763-82be-ef8587ee7628/267880ff-0fac-4763-82be-ef8587ee7628.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1003.755853] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b1d8b948-c036-441b-951d-2a3499736831 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.758799] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.766168] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1003.766168] env[61728]: value = "task-464669" [ 1003.766168] env[61728]: _type = "Task" [ 1003.766168] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.775901] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464669, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.062546] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e5d230a-9874-4bff-9d27-148276fe7070 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.074844] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7c311e-aaba-4e04-8f7f-ab7a0532ff7d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.093991] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "refresh_cache-378ef752-6d47-4861-baf1-064b2a180dcc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.094474] env[61728]: DEBUG nova.compute.manager [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Instance network_info: |[{"id": "f99497f4-e49f-4d6b-9bc8-75efe36beb4b", "address": "fa:16:3e:87:0a:68", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf99497f4-e4", "ovs_interfaceid": "f99497f4-e49f-4d6b-9bc8-75efe36beb4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1004.094857] env[61728]: DEBUG oslo_concurrency.lockutils [req-440250c4-5fbd-4b7b-96bc-e8a745a8db6b req-3b9bc160-a625-4478-adf2-9b7c281ec6da service nova] Acquired lock "refresh_cache-378ef752-6d47-4861-baf1-064b2a180dcc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.095070] env[61728]: DEBUG nova.network.neutron [req-440250c4-5fbd-4b7b-96bc-e8a745a8db6b req-3b9bc160-a625-4478-adf2-9b7c281ec6da service nova] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Refreshing network info cache for port f99497f4-e49f-4d6b-9bc8-75efe36beb4b {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1004.096523] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:0a:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c492f5cc-7ae0-4cab-823c-0d5dd8c60b26', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f99497f4-e49f-4d6b-9bc8-75efe36beb4b', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1004.106857] env[61728]: DEBUG oslo.service.loopingcall [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.137576] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1004.142987] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b4863b5-b14b-4f28-947b-8218023baeb9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.162797] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d24729-88e4-46f5-ab24-8cd6c9f9b3cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.189773] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1b9cdf-b050-42e2-b30d-4eed87e94455 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.197445] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1004.197445] env[61728]: value = "task-464670" [ 1004.197445] env[61728]: _type = "Task" [ 1004.197445] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.197445] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464667, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.218371] env[61728]: DEBUG nova.compute.provider_tree [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1004.226912] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464670, 'name': CreateVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.237626] env[61728]: DEBUG oslo_vmware.api [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464668, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.285358] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464669, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.650550] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464667, 'name': ReconfigVM_Task, 'duration_secs': 0.608008} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.657040] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f/0edd1f56-3e8e-4868-9c8e-a8caaf026e7f.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1004.657040] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-db6b5dac-e02b-4124-8ea5-43a7b989c33a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.663777] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1004.663777] env[61728]: value = "task-464671" [ 1004.663777] env[61728]: _type = "Task" [ 1004.663777] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.675536] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464671, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.714964] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464670, 'name': CreateVM_Task, 'duration_secs': 0.488855} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.714964] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1004.715865] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.715865] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.717218] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1004.717218] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-755128dd-4099-43cf-a6c7-36a7a05a0350 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.729882] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 1004.729882] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5296296f-2479-e434-5035-b3fd33187980" [ 1004.729882] env[61728]: _type = "Task" [ 1004.729882] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.736826] env[61728]: DEBUG oslo_vmware.api [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464668, 'name': ReconfigVM_Task, 'duration_secs': 0.856267} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.737286] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.737720] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Reconfigured VM to attach interface {{(pid=61728) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1004.741073] env[61728]: ERROR nova.scheduler.client.report [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [req-f3aefabe-e2e1-4e61-b6bc-5542a7ef27b2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f3aefabe-e2e1-4e61-b6bc-5542a7ef27b2"}]} [ 1004.750400] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5296296f-2479-e434-5035-b3fd33187980, 'name': SearchDatastore_Task, 'duration_secs': 0.013094} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.750793] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.751091] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1004.751359] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.751513] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.751740] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1004.752070] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf544e5e-f67a-45f3-9307-46fb0fe7113a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.762464] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1004.762788] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1004.764029] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7aba33b6-c501-4b94-8fe3-44994dfeff9f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.766742] env[61728]: DEBUG nova.scheduler.client.report [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1004.773521] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 1004.773521] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ff0801-52f7-9fbd-253a-2dc37565591d" [ 1004.773521] env[61728]: _type = "Task" [ 1004.773521] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.782676] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464669, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.612845} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.783406] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 267880ff-0fac-4763-82be-ef8587ee7628/267880ff-0fac-4763-82be-ef8587ee7628.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1004.786330] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1004.787812] env[61728]: DEBUG nova.scheduler.client.report [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1004.787992] env[61728]: DEBUG nova.compute.provider_tree [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1004.793239] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f877797b-48fa-4248-8ae9-7dc169f3175e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.795136] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ff0801-52f7-9fbd-253a-2dc37565591d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.801120] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1004.801120] env[61728]: value = "task-464672" [ 1004.801120] env[61728]: _type = "Task" [ 1004.801120] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.811200] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464672, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.812348] env[61728]: DEBUG nova.scheduler.client.report [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1004.842531] env[61728]: DEBUG nova.scheduler.client.report [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1004.921858] env[61728]: DEBUG nova.network.neutron [req-440250c4-5fbd-4b7b-96bc-e8a745a8db6b req-3b9bc160-a625-4478-adf2-9b7c281ec6da service nova] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Updated VIF entry in instance network info cache for port f99497f4-e49f-4d6b-9bc8-75efe36beb4b. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1004.922277] env[61728]: DEBUG nova.network.neutron [req-440250c4-5fbd-4b7b-96bc-e8a745a8db6b req-3b9bc160-a625-4478-adf2-9b7c281ec6da service nova] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Updating instance_info_cache with network_info: [{"id": "f99497f4-e49f-4d6b-9bc8-75efe36beb4b", "address": "fa:16:3e:87:0a:68", "network": {"id": "8f8ce06c-c9f6-49c2-91a4-b59d264dba01", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-978515845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5329f59d8bf43b2a855c4259387fccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf99497f4-e4", "ovs_interfaceid": "f99497f4-e49f-4d6b-9bc8-75efe36beb4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.111162] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c0d6f06-a9c2-4a98-b16e-2dd77d3984d1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.120197] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-083994fb-00d6-4649-9fb6-34fcf9e1ec61 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.158100] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57255c11-c756-4b4a-8a0e-8c5814167016 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.174986] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3742ff78-3a2b-457c-841b-892ff5dbac0c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.182169] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464671, 'name': Rename_Task, 'duration_secs': 0.189825} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.182835] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1005.183099] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6eb238bf-7500-42d8-a650-d1618bf23a00 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.193947] env[61728]: DEBUG nova.compute.provider_tree [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1005.197409] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1005.197409] env[61728]: value = "task-464673" [ 1005.197409] env[61728]: _type = "Task" [ 1005.197409] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.206173] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464673, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.245886] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2b0abfc8-0255-4c00-88c9-8217fb1a83e7 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-a70d1c37-e792-4168-b7e1-9418b8cb7818-d0b341ea-4a0b-46ff-89f2-f96d212d04d9" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.520s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.285666] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ff0801-52f7-9fbd-253a-2dc37565591d, 'name': SearchDatastore_Task, 'duration_secs': 0.022556} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.286546] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c541782-9582-400b-b22b-98441f208dd9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.292822] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 1005.292822] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522f25c2-188f-0674-c907-d39fd099b3d6" [ 1005.292822] env[61728]: _type = "Task" [ 1005.292822] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.305076] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522f25c2-188f-0674-c907-d39fd099b3d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.311988] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464672, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.178979} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.312401] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1005.313424] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581f756c-2a1d-4dba-8d61-1f988f93a3b7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.341464] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 267880ff-0fac-4763-82be-ef8587ee7628/267880ff-0fac-4763-82be-ef8587ee7628.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1005.341935] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f2c1229-2406-446c-a209-6703e9f9b769 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.362486] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1005.362486] env[61728]: value = "task-464674" [ 1005.362486] env[61728]: _type = "Task" [ 1005.362486] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.371894] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464674, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.428672] env[61728]: DEBUG oslo_concurrency.lockutils [req-440250c4-5fbd-4b7b-96bc-e8a745a8db6b req-3b9bc160-a625-4478-adf2-9b7c281ec6da service nova] Releasing lock "refresh_cache-378ef752-6d47-4861-baf1-064b2a180dcc" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.719397] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464673, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.736051] env[61728]: ERROR nova.scheduler.client.report [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] [req-217a9b92-36a3-4763-9390-a0b21c4d39a7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-217a9b92-36a3-4763-9390-a0b21c4d39a7"}]} [ 1005.759077] env[61728]: DEBUG nova.scheduler.client.report [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1005.775796] env[61728]: DEBUG nova.scheduler.client.report [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1005.776069] env[61728]: DEBUG nova.compute.provider_tree [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1005.798596] env[61728]: DEBUG nova.scheduler.client.report [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1005.809264] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522f25c2-188f-0674-c907-d39fd099b3d6, 'name': SearchDatastore_Task, 'duration_secs': 0.026216} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.810244] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.811156] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 378ef752-6d47-4861-baf1-064b2a180dcc/378ef752-6d47-4861-baf1-064b2a180dcc.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1005.811156] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1cd51002-c566-4d8b-a905-e0b5685d45e4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.819484] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 1005.819484] env[61728]: value = "task-464675" [ 1005.819484] env[61728]: _type = "Task" [ 1005.819484] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.826667] env[61728]: DEBUG nova.scheduler.client.report [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1005.841175] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Acquiring lock "9efdee89-4c88-4af3-af04-5281e8f6bc25" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.841175] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Lock "9efdee89-4c88-4af3-af04-5281e8f6bc25" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.842226] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464675, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.874191] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464674, 'name': ReconfigVM_Task, 'duration_secs': 0.408163} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.878132] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 267880ff-0fac-4763-82be-ef8587ee7628/267880ff-0fac-4763-82be-ef8587ee7628.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.878690] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e13098b-79e1-4177-b81d-808add68976e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.887008] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1005.887008] env[61728]: value = "task-464676" [ 1005.887008] env[61728]: _type = "Task" [ 1005.887008] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.897367] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464676, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.167955] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64839941-65d1-40b9-9a9f-b765b3bafa97 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.177142] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af19dfe4-b660-4bf0-ad6f-9fcdfe0a61c8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.221880] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0548029e-b793-4f98-987b-8812cd61f066 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.231273] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464673, 'name': PowerOnVM_Task, 'duration_secs': 0.619565} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.233659] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1006.233659] env[61728]: DEBUG nova.compute.manager [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.234585] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7de1444-c6ee-4e09-8a38-7b41031ac929 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.239868] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a55a20-2880-44c5-92ef-457959250972 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.262855] env[61728]: DEBUG nova.compute.provider_tree [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1006.331105] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464675, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.344220] env[61728]: DEBUG nova.compute.manager [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1006.394150] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.394424] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.399673] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464676, 'name': Rename_Task, 'duration_secs': 0.181644} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.399950] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1006.400635] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eed46dc3-5c78-4435-949d-4cd0d792f0a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.407527] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1006.407527] env[61728]: value = "task-464677" [ 1006.407527] env[61728]: _type = "Task" [ 1006.407527] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.415933] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464677, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.768366] env[61728]: INFO nova.compute.manager [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] bringing vm to original state: 'stopped' [ 1006.805702] env[61728]: DEBUG nova.scheduler.client.report [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 154 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1006.809470] env[61728]: DEBUG nova.compute.provider_tree [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 154 to 155 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1006.809470] env[61728]: DEBUG nova.compute.provider_tree [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1006.834880] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464675, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.52076} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.835216] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 378ef752-6d47-4861-baf1-064b2a180dcc/378ef752-6d47-4861-baf1-064b2a180dcc.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1006.835444] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1006.836185] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ed919ed-d277-43a9-98cd-587cdda76db0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.844701] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 1006.844701] env[61728]: value = "task-464678" [ 1006.844701] env[61728]: _type = "Task" [ 1006.844701] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.859496] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464678, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.872994] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.901056] env[61728]: DEBUG nova.compute.utils [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1006.923229] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464677, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.127533] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "interface-a70d1c37-e792-4168-b7e1-9418b8cb7818-d0b341ea-4a0b-46ff-89f2-f96d212d04d9" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.127820] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-a70d1c37-e792-4168-b7e1-9418b8cb7818-d0b341ea-4a0b-46ff-89f2-f96d212d04d9" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.313104] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 4.133s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.315609] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.290s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.315845] env[61728]: DEBUG nova.objects.instance [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lazy-loading 'resources' on Instance uuid 3bfefcfc-db97-4a9d-86cb-9fb1d8158863 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.335078] env[61728]: INFO nova.scheduler.client.report [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Deleted allocations for instance 7e8cde9c-890c-4f50-9ccb-46aa71220395 [ 1007.354285] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464678, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067764} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.354554] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1007.355362] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a364662-5aca-4c7d-86c6-e132ac63d3f0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.378427] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 378ef752-6d47-4861-baf1-064b2a180dcc/378ef752-6d47-4861-baf1-064b2a180dcc.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1007.378733] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf2d2622-3aa6-4456-a504-6f941a8eb030 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.401351] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 1007.401351] env[61728]: value = "task-464679" [ 1007.401351] env[61728]: _type = "Task" [ 1007.401351] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.405508] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.413224] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464679, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.421833] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464677, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.630664] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.630957] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.632259] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86beeb55-0387-4c86-8873-444fdcbf7431 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.658256] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbfc34ea-be94-444f-a217-5cf7a1637deb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.699689] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Reconfiguring VM to detach interface {{(pid=61728) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1007.700849] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8269696-8378-49cb-a3fe-0a87e9c15d5a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.727504] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 1007.727504] env[61728]: value = "task-464680" [ 1007.727504] env[61728]: _type = "Task" [ 1007.727504] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.736184] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464680, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.778067] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.778366] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.778564] env[61728]: DEBUG nova.compute.manager [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1007.779469] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e67115-9444-40ab-8871-e667a0c46ed5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.787535] env[61728]: DEBUG nova.compute.manager [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61728) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1007.789736] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1007.790856] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84a1503d-b8d1-4c9e-a491-fe256ed9cb16 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.797491] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1007.797491] env[61728]: value = "task-464681" [ 1007.797491] env[61728]: _type = "Task" [ 1007.797491] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.806275] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464681, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.842027] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2e7309e0-4e76-49b4-ae7c-db2220b6cfb9 tempest-ServerPasswordTestJSON-522592389 tempest-ServerPasswordTestJSON-522592389-project-member] Lock "7e8cde9c-890c-4f50-9ccb-46aa71220395" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.627s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.915919] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464679, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.928215] env[61728]: DEBUG oslo_vmware.api [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464677, 'name': PowerOnVM_Task, 'duration_secs': 1.21773} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.928865] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1007.929270] env[61728]: INFO nova.compute.manager [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Took 10.89 seconds to spawn the instance on the hypervisor. [ 1007.929652] env[61728]: DEBUG nova.compute.manager [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1007.930565] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f7d9ee-027a-4fd8-854e-ba97eafad718 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.098348] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd3daa1-7cc3-4ba4-b1aa-c718e95ce621 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.106851] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f5a37cf-72df-4fba-a9ee-a360d1e38b39 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.140757] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b588f8-f6f3-4a95-a261-cd9a617fbd7d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.149169] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995436ce-cf64-41a2-a156-61c1b8a3e03d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.163353] env[61728]: DEBUG nova.compute.provider_tree [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1008.238813] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464680, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.308021] env[61728]: DEBUG oslo_vmware.api [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464681, 'name': PowerOffVM_Task, 'duration_secs': 0.299924} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.308359] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1008.308543] env[61728]: DEBUG nova.compute.manager [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1008.309354] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b13d7d-5ad4-409c-aed9-1bd4eeb8ef5a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.412604] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464679, 'name': ReconfigVM_Task, 'duration_secs': 0.903868} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.413641] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 378ef752-6d47-4861-baf1-064b2a180dcc/378ef752-6d47-4861-baf1-064b2a180dcc.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1008.414481] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bc4f63b7-4f72-4b91-bc55-a6266b559e58 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.421801] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 1008.421801] env[61728]: value = "task-464682" [ 1008.421801] env[61728]: _type = "Task" [ 1008.421801] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.432967] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464682, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.453512] env[61728]: INFO nova.compute.manager [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Took 31.29 seconds to build instance. [ 1008.481318] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.484043] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.002s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.484043] env[61728]: INFO nova.compute.manager [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Attaching volume 9e2549ce-9f21-4a53-af77-978682bfcbbe to /dev/sdb [ 1008.528156] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd515368-a528-4ffa-93c5-76bc9ff700c0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.536577] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f1adc4-a151-4d6a-9623-d3707169e6f0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.556292] env[61728]: DEBUG nova.virt.block_device [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Updating existing volume attachment record: 2b334db2-eb58-4539-837a-0ad6ccf3a726 {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1008.709461] env[61728]: DEBUG nova.scheduler.client.report [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 155 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1008.709461] env[61728]: DEBUG nova.compute.provider_tree [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 155 to 156 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1008.709461] env[61728]: DEBUG nova.compute.provider_tree [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1008.742314] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464680, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.824654] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.046s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.939210] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464682, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.960784] env[61728]: DEBUG oslo_concurrency.lockutils [None req-dc13b2e0-de49-4512-9206-02d91e3d86c5 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "267880ff-0fac-4763-82be-ef8587ee7628" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.814s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.217468] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.902s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.219741] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 6.772s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.245396] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464680, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.253163] env[61728]: INFO nova.scheduler.client.report [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Deleted allocations for instance 3bfefcfc-db97-4a9d-86cb-9fb1d8158863 [ 1009.338447] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.435454] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464682, 'name': Rename_Task, 'duration_secs': 0.983611} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.436381] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1009.436381] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aab06238-debc-4dd6-8f7d-b4d51b55561e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.443767] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 1009.443767] env[61728]: value = "task-464686" [ 1009.443767] env[61728]: _type = "Task" [ 1009.443767] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.459743] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464686, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.724926] env[61728]: DEBUG nova.objects.instance [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'migration_context' on Instance uuid f9f07464-9f43-43fd-8895-2673861747fa {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.745315] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464680, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.762922] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a413799-79f9-4aca-84ed-de7d444553a2 tempest-ServersNegativeTestJSON-1621560623 tempest-ServersNegativeTestJSON-1621560623-project-member] Lock "3bfefcfc-db97-4a9d-86cb-9fb1d8158863" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.199s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.956471] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464686, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.053145] env[61728]: DEBUG nova.compute.manager [req-384e2816-daae-40cf-b108-93b0cd9e5459 req-a50b70af-162b-4aa9-868b-4ca7d1a288c2 service nova] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Received event network-changed-eb281159-3780-4ea6-b463-cd8d839be907 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1010.053352] env[61728]: DEBUG nova.compute.manager [req-384e2816-daae-40cf-b108-93b0cd9e5459 req-a50b70af-162b-4aa9-868b-4ca7d1a288c2 service nova] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Refreshing instance network info cache due to event network-changed-eb281159-3780-4ea6-b463-cd8d839be907. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1010.053609] env[61728]: DEBUG oslo_concurrency.lockutils [req-384e2816-daae-40cf-b108-93b0cd9e5459 req-a50b70af-162b-4aa9-868b-4ca7d1a288c2 service nova] Acquiring lock "refresh_cache-267880ff-0fac-4763-82be-ef8587ee7628" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.053727] env[61728]: DEBUG oslo_concurrency.lockutils [req-384e2816-daae-40cf-b108-93b0cd9e5459 req-a50b70af-162b-4aa9-868b-4ca7d1a288c2 service nova] Acquired lock "refresh_cache-267880ff-0fac-4763-82be-ef8587ee7628" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.053894] env[61728]: DEBUG nova.network.neutron [req-384e2816-daae-40cf-b108-93b0cd9e5459 req-a50b70af-162b-4aa9-868b-4ca7d1a288c2 service nova] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Refreshing network info cache for port eb281159-3780-4ea6-b463-cd8d839be907 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1010.248635] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464680, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.363226] env[61728]: DEBUG oslo_concurrency.lockutils [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.363596] env[61728]: DEBUG oslo_concurrency.lockutils [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.363946] env[61728]: DEBUG oslo_concurrency.lockutils [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.364354] env[61728]: DEBUG oslo_concurrency.lockutils [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.364593] env[61728]: DEBUG oslo_concurrency.lockutils [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.373804] env[61728]: INFO nova.compute.manager [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Terminating instance [ 1010.376818] env[61728]: DEBUG nova.compute.manager [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1010.377068] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1010.378567] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b557aa2-90e1-4d86-8e41-5545f23e22bb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.391607] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1010.391924] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d5af45a5-9e62-4788-90c8-d99aeb075a66 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.459773] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464686, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.469509] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1010.469509] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1010.469509] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleting the datastore file [datastore1] 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.469509] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af043770-b53f-4250-80f4-ff907b748975 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.480064] env[61728]: DEBUG oslo_vmware.api [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1010.480064] env[61728]: value = "task-464688" [ 1010.480064] env[61728]: _type = "Task" [ 1010.480064] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.496168] env[61728]: DEBUG oslo_vmware.api [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464688, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.554110] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c49b328-13f0-4c1b-b97a-bf530c675eb4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.566603] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed3fce9-90a4-4506-89c0-a791251938f3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.604316] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b9dc2f-d1eb-4b68-8eeb-4a37f317b841 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.611465] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d7a69d7-6aab-448c-8582-5096067efc7a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.625338] env[61728]: DEBUG nova.compute.provider_tree [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.743882] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464680, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.958496] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464686, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.999060] env[61728]: DEBUG oslo_vmware.api [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464688, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249474} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.999060] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1010.999060] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1010.999060] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1010.999060] env[61728]: INFO nova.compute.manager [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1010.999060] env[61728]: DEBUG oslo.service.loopingcall [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.999060] env[61728]: DEBUG nova.compute.manager [-] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1010.999060] env[61728]: DEBUG nova.network.neutron [-] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1011.134544] env[61728]: DEBUG nova.scheduler.client.report [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1011.240905] env[61728]: DEBUG nova.network.neutron [req-384e2816-daae-40cf-b108-93b0cd9e5459 req-a50b70af-162b-4aa9-868b-4ca7d1a288c2 service nova] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Updated VIF entry in instance network info cache for port eb281159-3780-4ea6-b463-cd8d839be907. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1011.240905] env[61728]: DEBUG nova.network.neutron [req-384e2816-daae-40cf-b108-93b0cd9e5459 req-a50b70af-162b-4aa9-868b-4ca7d1a288c2 service nova] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Updating instance_info_cache with network_info: [{"id": "eb281159-3780-4ea6-b463-cd8d839be907", "address": "fa:16:3e:ff:b3:0b", "network": {"id": "22ad4b68-2895-4651-803f-44b8031a6292", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-489227259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "276942c8cf8a42729a541de096f69a0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb281159-37", "ovs_interfaceid": "eb281159-3780-4ea6-b463-cd8d839be907", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.250469] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464680, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.258764] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Acquiring lock "5f248cf6-31a9-49ef-a8a8-218987664af7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.259073] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Lock "5f248cf6-31a9-49ef-a8a8-218987664af7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.378362] env[61728]: DEBUG nova.compute.manager [req-c85f3487-9110-41d4-80ec-9bd705ab9170 req-ec4c2bd9-2f3f-4ee9-aeff-590de98578f3 service nova] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Received event network-vif-deleted-18a00538-a062-4136-b2d4-2b3b3f783502 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1011.378362] env[61728]: INFO nova.compute.manager [req-c85f3487-9110-41d4-80ec-9bd705ab9170 req-ec4c2bd9-2f3f-4ee9-aeff-590de98578f3 service nova] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Neutron deleted interface 18a00538-a062-4136-b2d4-2b3b3f783502; detaching it from the instance and deleting it from the info cache [ 1011.378362] env[61728]: DEBUG nova.network.neutron [req-c85f3487-9110-41d4-80ec-9bd705ab9170 req-ec4c2bd9-2f3f-4ee9-aeff-590de98578f3 service nova] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.464806] env[61728]: DEBUG oslo_vmware.api [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464686, 'name': PowerOnVM_Task, 'duration_secs': 1.6117} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.465282] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1011.465662] env[61728]: INFO nova.compute.manager [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Took 10.18 seconds to spawn the instance on the hypervisor. [ 1011.465989] env[61728]: DEBUG nova.compute.manager [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1011.467296] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b90c87e-0ddc-4b3f-b8f9-450de8b5a6ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.745661] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464680, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.747400] env[61728]: DEBUG oslo_concurrency.lockutils [req-384e2816-daae-40cf-b108-93b0cd9e5459 req-a50b70af-162b-4aa9-868b-4ca7d1a288c2 service nova] Releasing lock "refresh_cache-267880ff-0fac-4763-82be-ef8587ee7628" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.762288] env[61728]: DEBUG nova.compute.manager [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1011.847832] env[61728]: DEBUG nova.network.neutron [-] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.881229] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ea72698-36d1-4b0e-a80b-99f1e89a7d4f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.891298] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b59e85-94af-49ee-b318-3b1dc9535bed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.925434] env[61728]: DEBUG nova.compute.manager [req-c85f3487-9110-41d4-80ec-9bd705ab9170 req-ec4c2bd9-2f3f-4ee9-aeff-590de98578f3 service nova] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Detach interface failed, port_id=18a00538-a062-4136-b2d4-2b3b3f783502, reason: Instance 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1011.989480] env[61728]: INFO nova.compute.manager [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Took 33.59 seconds to build instance. [ 1012.151318] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.931s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.157766] env[61728]: DEBUG oslo_concurrency.lockutils [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.457s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.158233] env[61728]: DEBUG nova.objects.instance [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lazy-loading 'resources' on Instance uuid 172129ff-136e-4855-8659-3f1ac88dcbd2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1012.246189] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464680, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.289055] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.351171] env[61728]: INFO nova.compute.manager [-] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Took 1.35 seconds to deallocate network for instance. [ 1012.494772] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bad15720-8e46-4007-81e9-1546ba8b99a7 tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "378ef752-6d47-4861-baf1-064b2a180dcc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.107s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.745981] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464680, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.858553] env[61728]: DEBUG oslo_concurrency.lockutils [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.947847] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a949d43-2898-42a2-a20f-99e5941a364b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.957256] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1cc0fe2-4ad6-4d48-819e-c3f4ae5008e9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.992341] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd53704-9830-4007-962c-08fae2eb80bf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.001029] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-868d8aea-f570-4d63-b337-8dc7e0ccc996 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.014245] env[61728]: DEBUG nova.compute.provider_tree [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.121035] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1013.121356] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122182', 'volume_id': '9e2549ce-9f21-4a53-af77-978682bfcbbe', 'name': 'volume-9e2549ce-9f21-4a53-af77-978682bfcbbe', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a', 'attached_at': '', 'detached_at': '', 'volume_id': '9e2549ce-9f21-4a53-af77-978682bfcbbe', 'serial': '9e2549ce-9f21-4a53-af77-978682bfcbbe'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1013.122277] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2101be1-9450-47f0-ac38-da1de6e8d318 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.140559] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d70308e-8e6d-419d-98fa-fcadb80b2a75 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.166657] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] volume-9e2549ce-9f21-4a53-af77-978682bfcbbe/volume-9e2549ce-9f21-4a53-af77-978682bfcbbe.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.166975] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-974005a7-c014-4f3b-9f7b-e11d22c10899 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.190692] env[61728]: DEBUG oslo_vmware.api [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1013.190692] env[61728]: value = "task-464690" [ 1013.190692] env[61728]: _type = "Task" [ 1013.190692] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.199168] env[61728]: DEBUG oslo_vmware.api [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464690, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.246477] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464680, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.518261] env[61728]: DEBUG nova.scheduler.client.report [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.613729] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "378ef752-6d47-4861-baf1-064b2a180dcc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.613729] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "378ef752-6d47-4861-baf1-064b2a180dcc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.613729] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "378ef752-6d47-4861-baf1-064b2a180dcc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.613729] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "378ef752-6d47-4861-baf1-064b2a180dcc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.613729] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "378ef752-6d47-4861-baf1-064b2a180dcc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.617349] env[61728]: INFO nova.compute.manager [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Terminating instance [ 1013.619778] env[61728]: DEBUG nova.compute.manager [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1013.620016] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1013.621099] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a42546-30a1-4986-86b2-f8a66e00a01b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.634472] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1013.634747] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-de1ea470-c9b4-4008-9006-a8167a0bd3c4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.641532] env[61728]: DEBUG oslo_vmware.api [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 1013.641532] env[61728]: value = "task-464691" [ 1013.641532] env[61728]: _type = "Task" [ 1013.641532] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.649480] env[61728]: DEBUG oslo_vmware.api [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464691, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.701578] env[61728]: DEBUG oslo_vmware.api [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464690, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.718042] env[61728]: INFO nova.compute.manager [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Swapping old allocation on dict_keys(['e7ceb92f-072b-409e-b888-6fe0676b32f1']) held by migration 258316b5-7ba1-41d5-9080-956cb4a7720c for instance [ 1013.748275] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464680, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.756238] env[61728]: DEBUG nova.scheduler.client.report [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Overwriting current allocation {'allocations': {'e7ceb92f-072b-409e-b888-6fe0676b32f1': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 157}}, 'project_id': '49c64edaa37f41d38aba7bd5b1d0b47f', 'user_id': '31a2286aad9a40da9b16353541a2780f', 'consumer_generation': 1} on consumer f9f07464-9f43-43fd-8895-2673861747fa {{(pid=61728) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1013.844733] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.844930] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.845139] env[61728]: DEBUG nova.network.neutron [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1014.025781] env[61728]: DEBUG oslo_concurrency.lockutils [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.868s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.028980] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.270s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.032135] env[61728]: INFO nova.compute.claims [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1014.041990] env[61728]: INFO nova.scheduler.client.report [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleted allocations for instance 172129ff-136e-4855-8659-3f1ac88dcbd2 [ 1014.152975] env[61728]: DEBUG oslo_vmware.api [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464691, 'name': PowerOffVM_Task, 'duration_secs': 0.184701} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.153354] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1014.153620] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1014.153965] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-19ad9a35-13a5-4164-a3ea-2209d1b5615e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.206362] env[61728]: DEBUG oslo_vmware.api [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464690, 'name': ReconfigVM_Task, 'duration_secs': 0.579633} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.206362] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Reconfigured VM instance instance-0000005c to attach disk [datastore1] volume-9e2549ce-9f21-4a53-af77-978682bfcbbe/volume-9e2549ce-9f21-4a53-af77-978682bfcbbe.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.211222] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71f6e82d-d892-46da-a512-36185f2091fc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.227635] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1014.227779] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1014.228023] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleting the datastore file [datastore1] 378ef752-6d47-4861-baf1-064b2a180dcc {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1014.229362] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-acb5c272-95f7-4600-95fc-db0dcdcc357e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.231278] env[61728]: DEBUG oslo_vmware.api [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1014.231278] env[61728]: value = "task-464693" [ 1014.231278] env[61728]: _type = "Task" [ 1014.231278] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.237975] env[61728]: DEBUG oslo_vmware.api [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for the task: (returnval){ [ 1014.237975] env[61728]: value = "task-464694" [ 1014.237975] env[61728]: _type = "Task" [ 1014.237975] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.245342] env[61728]: DEBUG oslo_vmware.api [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464693, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.258226] env[61728]: DEBUG oslo_vmware.api [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464694, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.261391] env[61728]: DEBUG oslo_vmware.api [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464680, 'name': ReconfigVM_Task, 'duration_secs': 6.472342} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.261776] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.262142] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Reconfigured VM to detach interface {{(pid=61728) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1014.553055] env[61728]: DEBUG oslo_concurrency.lockutils [None req-557bf851-2e14-4c4a-a041-d9ae70835d03 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "172129ff-136e-4855-8659-3f1ac88dcbd2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.770s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.765723] env[61728]: DEBUG oslo_vmware.api [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Task: {'id': task-464694, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127126} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.765723] env[61728]: DEBUG oslo_vmware.api [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464693, 'name': ReconfigVM_Task, 'duration_secs': 0.151068} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.765723] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.765723] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1014.765723] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1014.765723] env[61728]: INFO nova.compute.manager [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1014.765723] env[61728]: DEBUG oslo.service.loopingcall [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.765723] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122182', 'volume_id': '9e2549ce-9f21-4a53-af77-978682bfcbbe', 'name': 'volume-9e2549ce-9f21-4a53-af77-978682bfcbbe', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a', 'attached_at': '', 'detached_at': '', 'volume_id': '9e2549ce-9f21-4a53-af77-978682bfcbbe', 'serial': '9e2549ce-9f21-4a53-af77-978682bfcbbe'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1014.765723] env[61728]: DEBUG nova.compute.manager [-] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1014.765723] env[61728]: DEBUG nova.network.neutron [-] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1014.767708] env[61728]: DEBUG nova.network.neutron [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance_info_cache with network_info: [{"id": "83657468-5cac-4258-84f4-b105395d9cfa", "address": "fa:16:3e:55:1a:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83657468-5c", "ovs_interfaceid": "83657468-5cac-4258-84f4-b105395d9cfa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.081170] env[61728]: DEBUG nova.compute.manager [req-ee503f8f-5a68-4bcc-bad9-c13c528797ba req-22d1b54c-7dbc-4071-83bc-6fb00b266097 service nova] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Received event network-vif-deleted-f99497f4-e49f-4d6b-9bc8-75efe36beb4b {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1015.081599] env[61728]: INFO nova.compute.manager [req-ee503f8f-5a68-4bcc-bad9-c13c528797ba req-22d1b54c-7dbc-4071-83bc-6fb00b266097 service nova] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Neutron deleted interface f99497f4-e49f-4d6b-9bc8-75efe36beb4b; detaching it from the instance and deleting it from the info cache [ 1015.081924] env[61728]: DEBUG nova.network.neutron [req-ee503f8f-5a68-4bcc-bad9-c13c528797ba req-22d1b54c-7dbc-4071-83bc-6fb00b266097 service nova] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.273500] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "refresh_cache-f9f07464-9f43-43fd-8895-2673861747fa" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.275730] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20074127-7603-42fe-b449-78bf964a2a76 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.278945] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8865f12-e322-41c8-9441-db2f995e9ff2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.289038] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a483fa-92d0-4912-9d9e-7f65298922e8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.292714] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb1a158-3d93-4d4d-bbe4-adfd095d0fb3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.335531] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ada057-7bfe-4d3b-b916-39fe84f0dbb4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.349689] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d0fb14-f61d-4fea-a1ca-72042d7eac01 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.365280] env[61728]: DEBUG nova.compute.provider_tree [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.535986] env[61728]: DEBUG nova.network.neutron [-] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.584921] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ac86919-be2c-4e48-93b9-ae4240a2d97c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.594670] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd34c2bf-68d9-4a68-aff9-9f5f46e01616 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.626094] env[61728]: DEBUG nova.compute.manager [req-ee503f8f-5a68-4bcc-bad9-c13c528797ba req-22d1b54c-7dbc-4071-83bc-6fb00b266097 service nova] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Detach interface failed, port_id=f99497f4-e49f-4d6b-9bc8-75efe36beb4b, reason: Instance 378ef752-6d47-4861-baf1-064b2a180dcc could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1015.844120] env[61728]: DEBUG nova.objects.instance [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lazy-loading 'flavor' on Instance uuid 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.844120] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.844120] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.844120] env[61728]: DEBUG nova.network.neutron [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1015.868916] env[61728]: DEBUG nova.scheduler.client.report [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.040466] env[61728]: INFO nova.compute.manager [-] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Took 1.28 seconds to deallocate network for instance. [ 1016.307877] env[61728]: DEBUG nova.compute.manager [req-521c26c6-3ae0-4608-8346-09b4b787142b req-4883b575-5e74-4428-bb20-dc4a03122bff service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Received event network-changed-c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.307963] env[61728]: DEBUG nova.compute.manager [req-521c26c6-3ae0-4608-8346-09b4b787142b req-4883b575-5e74-4428-bb20-dc4a03122bff service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Refreshing instance network info cache due to event network-changed-c8407eac-4337-4ae3-a071-08f48bf87248. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1016.308197] env[61728]: DEBUG oslo_concurrency.lockutils [req-521c26c6-3ae0-4608-8346-09b4b787142b req-4883b575-5e74-4428-bb20-dc4a03122bff service nova] Acquiring lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.332250] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.332250] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.349753] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9ab15a64-facd-4c2e-bb1f-12962eafbdb6 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.868s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.373819] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.377017] env[61728]: DEBUG nova.compute.manager [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1016.377451] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.504s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.379415] env[61728]: INFO nova.compute.claims [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1016.417357] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1016.417357] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-89a24e19-acbe-4316-9d64-0b06afe51567 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.426372] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1016.426372] env[61728]: value = "task-464695" [ 1016.426372] env[61728]: _type = "Task" [ 1016.426372] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.441567] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464695, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.547938] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.588807] env[61728]: INFO nova.network.neutron [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Port d0b341ea-4a0b-46ff-89f2-f96d212d04d9 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1016.589251] env[61728]: DEBUG nova.network.neutron [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updating instance_info_cache with network_info: [{"id": "c8407eac-4337-4ae3-a071-08f48bf87248", "address": "fa:16:3e:2d:a6:d6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8407eac-43", "ovs_interfaceid": "c8407eac-4337-4ae3-a071-08f48bf87248", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.794808] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "6c154517-68e8-4e70-86eb-cf53f6f812d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.795180] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "6c154517-68e8-4e70-86eb-cf53f6f812d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.835733] env[61728]: DEBUG nova.compute.manager [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1016.884086] env[61728]: DEBUG nova.compute.utils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1016.887803] env[61728]: DEBUG nova.compute.manager [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1016.888041] env[61728]: DEBUG nova.network.neutron [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1016.930253] env[61728]: DEBUG nova.policy [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54c30aeeea3a4c1b8ed34430361c2a4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a02bd852c564acd99e12fd17279f028', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1016.938134] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464695, 'name': PowerOffVM_Task, 'duration_secs': 0.169077} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.938535] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1016.939291] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1016.939593] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1016.939854] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.940086] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1016.940315] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.940572] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1016.940864] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1016.941111] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1016.941363] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1016.941614] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1016.941900] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1016.947260] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21e8088e-5b7f-472d-afda-af4c90b79d17 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.964361] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1016.964361] env[61728]: value = "task-464696" [ 1016.964361] env[61728]: _type = "Task" [ 1016.964361] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.975924] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464696, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.984784] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "interface-705c8838-f605-46fa-b036-ec212f673f66-d0b341ea-4a0b-46ff-89f2-f96d212d04d9" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.985251] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-705c8838-f605-46fa-b036-ec212f673f66-d0b341ea-4a0b-46ff-89f2-f96d212d04d9" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.985762] env[61728]: DEBUG nova.objects.instance [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'flavor' on Instance uuid 705c8838-f605-46fa-b036-ec212f673f66 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.049198] env[61728]: INFO nova.compute.manager [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Rescuing [ 1017.049387] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.049584] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.049941] env[61728]: DEBUG nova.network.neutron [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1017.092738] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.096843] env[61728]: DEBUG oslo_concurrency.lockutils [req-521c26c6-3ae0-4608-8346-09b4b787142b req-4883b575-5e74-4428-bb20-dc4a03122bff service nova] Acquired lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.098416] env[61728]: DEBUG nova.network.neutron [req-521c26c6-3ae0-4608-8346-09b4b787142b req-4883b575-5e74-4428-bb20-dc4a03122bff service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Refreshing network info cache for port c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1017.211922] env[61728]: DEBUG nova.network.neutron [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Successfully created port: 1db8e5be-43b2-4fa8-954e-dce8e21ff0d4 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1017.298293] env[61728]: DEBUG nova.compute.manager [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1017.359890] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.388421] env[61728]: DEBUG nova.compute.manager [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1017.488040] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464696, 'name': ReconfigVM_Task, 'duration_secs': 0.152288} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.496361] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907cbfd9-0ab8-49a5-a1ec-0ba873515423 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.521376] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1017.521376] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1017.521376] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1017.521617] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1017.521742] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1017.521886] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1017.522113] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1017.522285] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1017.522462] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1017.522633] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1017.522902] env[61728]: DEBUG nova.virt.hardware [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1017.528077] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96ce5280-add4-4b54-8a92-f5246369e018 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.533974] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1017.533974] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a0e994-baad-5b27-5735-e1e13839218a" [ 1017.533974] env[61728]: _type = "Task" [ 1017.533974] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.541623] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a0e994-baad-5b27-5735-e1e13839218a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.601336] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a726dda-8698-4ea0-9c11-46aba19c0614 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-a70d1c37-e792-4168-b7e1-9418b8cb7818-d0b341ea-4a0b-46ff-89f2-f96d212d04d9" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.473s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.637196] env[61728]: DEBUG nova.objects.instance [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'pci_requests' on Instance uuid 705c8838-f605-46fa-b036-ec212f673f66 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.699105] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f406606-debf-40e8-8f0c-85a5bce66a82 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.711304] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fbb2eed-0243-44bc-8250-16d167941b96 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.747265] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7ef94e-ae54-4e99-b295-1822fb51d9e2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.754683] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0cb24c0-5f35-44d4-a254-683150b0d0d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.768330] env[61728]: DEBUG nova.compute.provider_tree [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.819604] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.823783] env[61728]: DEBUG nova.network.neutron [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Updating instance_info_cache with network_info: [{"id": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "address": "fa:16:3e:d1:31:88", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67077c45-d8", "ovs_interfaceid": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.883114] env[61728]: DEBUG nova.network.neutron [req-521c26c6-3ae0-4608-8346-09b4b787142b req-4883b575-5e74-4428-bb20-dc4a03122bff service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updated VIF entry in instance network info cache for port c8407eac-4337-4ae3-a071-08f48bf87248. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1017.883567] env[61728]: DEBUG nova.network.neutron [req-521c26c6-3ae0-4608-8346-09b4b787142b req-4883b575-5e74-4428-bb20-dc4a03122bff service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updating instance_info_cache with network_info: [{"id": "c8407eac-4337-4ae3-a071-08f48bf87248", "address": "fa:16:3e:2d:a6:d6", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8407eac-43", "ovs_interfaceid": "c8407eac-4337-4ae3-a071-08f48bf87248", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.044446] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a0e994-baad-5b27-5735-e1e13839218a, 'name': SearchDatastore_Task, 'duration_secs': 0.031273} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.049834] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Reconfiguring VM instance instance-00000051 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1018.050156] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3755c84b-7c6b-400e-a637-5698906b4eb6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.068414] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1018.068414] env[61728]: value = "task-464697" [ 1018.068414] env[61728]: _type = "Task" [ 1018.068414] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.076758] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464697, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.144291] env[61728]: DEBUG nova.objects.base [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Object Instance<705c8838-f605-46fa-b036-ec212f673f66> lazy-loaded attributes: flavor,pci_requests {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1018.144536] env[61728]: DEBUG nova.network.neutron [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1018.236253] env[61728]: DEBUG nova.policy [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e51efcf23db4d5496edbc3b7104b4e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32bb35e1dfce40e48be08bb568d3f2b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1018.272652] env[61728]: DEBUG nova.scheduler.client.report [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1018.326796] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.386488] env[61728]: DEBUG oslo_concurrency.lockutils [req-521c26c6-3ae0-4608-8346-09b4b787142b req-4883b575-5e74-4428-bb20-dc4a03122bff service nova] Releasing lock "refresh_cache-a70d1c37-e792-4168-b7e1-9418b8cb7818" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.400276] env[61728]: DEBUG nova.compute.manager [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1018.405656] env[61728]: DEBUG nova.compute.manager [req-6eb76848-47e7-4e7d-b7e3-4cb7b8f3b89e req-b49b125f-cfd0-491d-89f1-bcd8fc5dbc8c service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Received event network-changed-ec0ae08d-62b0-4506-bc0b-65e9152dd5e8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.405895] env[61728]: DEBUG nova.compute.manager [req-6eb76848-47e7-4e7d-b7e3-4cb7b8f3b89e req-b49b125f-cfd0-491d-89f1-bcd8fc5dbc8c service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Refreshing instance network info cache due to event network-changed-ec0ae08d-62b0-4506-bc0b-65e9152dd5e8. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1018.406172] env[61728]: DEBUG oslo_concurrency.lockutils [req-6eb76848-47e7-4e7d-b7e3-4cb7b8f3b89e req-b49b125f-cfd0-491d-89f1-bcd8fc5dbc8c service nova] Acquiring lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.406364] env[61728]: DEBUG oslo_concurrency.lockutils [req-6eb76848-47e7-4e7d-b7e3-4cb7b8f3b89e req-b49b125f-cfd0-491d-89f1-bcd8fc5dbc8c service nova] Acquired lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.406580] env[61728]: DEBUG nova.network.neutron [req-6eb76848-47e7-4e7d-b7e3-4cb7b8f3b89e req-b49b125f-cfd0-491d-89f1-bcd8fc5dbc8c service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Refreshing network info cache for port ec0ae08d-62b0-4506-bc0b-65e9152dd5e8 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1018.427293] env[61728]: DEBUG nova.virt.hardware [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1018.427561] env[61728]: DEBUG nova.virt.hardware [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1018.427735] env[61728]: DEBUG nova.virt.hardware [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1018.427925] env[61728]: DEBUG nova.virt.hardware [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1018.428090] env[61728]: DEBUG nova.virt.hardware [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1018.428246] env[61728]: DEBUG nova.virt.hardware [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1018.428450] env[61728]: DEBUG nova.virt.hardware [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1018.428615] env[61728]: DEBUG nova.virt.hardware [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1018.428791] env[61728]: DEBUG nova.virt.hardware [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1018.428960] env[61728]: DEBUG nova.virt.hardware [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1018.429407] env[61728]: DEBUG nova.virt.hardware [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1018.430803] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ee5af8-b97f-4a4d-b679-2e27d8a320e5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.439284] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a5d96b-07fe-41aa-811a-e7f9dd4967c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.578486] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464697, 'name': ReconfigVM_Task, 'duration_secs': 0.223725} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.578954] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Reconfigured VM instance instance-00000051 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1018.579855] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1885c88-b110-4ff0-91ba-44819c7a6c47 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.606697] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] f9f07464-9f43-43fd-8895-2673861747fa/f9f07464-9f43-43fd-8895-2673861747fa.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.608384] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c62abae-82d5-457b-9a69-df00bdb729db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.622717] env[61728]: DEBUG nova.compute.manager [req-75454880-f9c0-491b-89e1-68f73887f042 req-ab660d83-4be0-496f-848a-1f783186263f service nova] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Received event network-vif-plugged-1db8e5be-43b2-4fa8-954e-dce8e21ff0d4 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.623088] env[61728]: DEBUG oslo_concurrency.lockutils [req-75454880-f9c0-491b-89e1-68f73887f042 req-ab660d83-4be0-496f-848a-1f783186263f service nova] Acquiring lock "bb7b38d5-53c7-4c05-96f8-9121171d3c46-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.623417] env[61728]: DEBUG oslo_concurrency.lockutils [req-75454880-f9c0-491b-89e1-68f73887f042 req-ab660d83-4be0-496f-848a-1f783186263f service nova] Lock "bb7b38d5-53c7-4c05-96f8-9121171d3c46-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.623719] env[61728]: DEBUG oslo_concurrency.lockutils [req-75454880-f9c0-491b-89e1-68f73887f042 req-ab660d83-4be0-496f-848a-1f783186263f service nova] Lock "bb7b38d5-53c7-4c05-96f8-9121171d3c46-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.624016] env[61728]: DEBUG nova.compute.manager [req-75454880-f9c0-491b-89e1-68f73887f042 req-ab660d83-4be0-496f-848a-1f783186263f service nova] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] No waiting events found dispatching network-vif-plugged-1db8e5be-43b2-4fa8-954e-dce8e21ff0d4 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1018.625101] env[61728]: WARNING nova.compute.manager [req-75454880-f9c0-491b-89e1-68f73887f042 req-ab660d83-4be0-496f-848a-1f783186263f service nova] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Received unexpected event network-vif-plugged-1db8e5be-43b2-4fa8-954e-dce8e21ff0d4 for instance with vm_state building and task_state spawning. [ 1018.631038] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1018.631038] env[61728]: value = "task-464698" [ 1018.631038] env[61728]: _type = "Task" [ 1018.631038] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.638905] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464698, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.681686] env[61728]: DEBUG nova.network.neutron [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Successfully updated port: 1db8e5be-43b2-4fa8-954e-dce8e21ff0d4 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1018.780035] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.780396] env[61728]: DEBUG nova.compute.manager [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1018.783440] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 9.445s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.783677] env[61728]: DEBUG nova.objects.instance [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61728) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1018.868401] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1018.868637] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e3b105e1-9ff1-474b-bdbd-abbd76670a94 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.877934] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1018.877934] env[61728]: value = "task-464699" [ 1018.877934] env[61728]: _type = "Task" [ 1018.877934] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.887657] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464699, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.137606] env[61728]: DEBUG nova.network.neutron [req-6eb76848-47e7-4e7d-b7e3-4cb7b8f3b89e req-b49b125f-cfd0-491d-89f1-bcd8fc5dbc8c service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Updated VIF entry in instance network info cache for port ec0ae08d-62b0-4506-bc0b-65e9152dd5e8. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1019.138036] env[61728]: DEBUG nova.network.neutron [req-6eb76848-47e7-4e7d-b7e3-4cb7b8f3b89e req-b49b125f-cfd0-491d-89f1-bcd8fc5dbc8c service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Updating instance_info_cache with network_info: [{"id": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "address": "fa:16:3e:57:58:2b", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec0ae08d-62", "ovs_interfaceid": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.142557] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464698, 'name': ReconfigVM_Task, 'duration_secs': 0.268845} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.143204] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Reconfigured VM instance instance-00000051 to attach disk [datastore1] f9f07464-9f43-43fd-8895-2673861747fa/f9f07464-9f43-43fd-8895-2673861747fa.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.143878] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d9c2de-199f-4658-bb16-82fdc1bad023 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.165349] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a709c5c6-07db-40f3-ab18-16411c59957c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.187755] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "refresh_cache-bb7b38d5-53c7-4c05-96f8-9121171d3c46" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.187932] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "refresh_cache-bb7b38d5-53c7-4c05-96f8-9121171d3c46" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.188050] env[61728]: DEBUG nova.network.neutron [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1019.189635] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917d58ca-d509-40a2-aeb7-8566803e8066 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.211655] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91b66a5f-82c7-4b81-8aaa-0c40035980f1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.219037] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1019.219294] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9850cac0-d1b9-460b-930c-1a71a7bda16a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.225894] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1019.225894] env[61728]: value = "task-464700" [ 1019.225894] env[61728]: _type = "Task" [ 1019.225894] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.233325] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464700, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.290455] env[61728]: DEBUG nova.compute.utils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1019.292201] env[61728]: DEBUG nova.compute.manager [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1019.292397] env[61728]: DEBUG nova.network.neutron [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1019.336406] env[61728]: DEBUG nova.policy [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0fe70ce1eff04a90ae8bbfa39c8e8f32', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9f255ab8e7a40cb86b7bdb1f3aadf2c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1019.385723] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464699, 'name': PowerOffVM_Task, 'duration_secs': 0.196658} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.386051] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1019.386832] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6fc034c-9f10-4152-8603-3d5c540c427e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.407421] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31c4a06-8c19-4778-8629-76555eafcf6a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.441786] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1019.441786] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d1e8f30-539f-40e4-bb3e-7e0210653063 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.448495] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1019.448495] env[61728]: value = "task-464701" [ 1019.448495] env[61728]: _type = "Task" [ 1019.448495] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.456434] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464701, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.598868] env[61728]: DEBUG nova.network.neutron [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Successfully created port: 216193d1-1332-49ba-8702-dacc02486ec1 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1019.643976] env[61728]: DEBUG oslo_concurrency.lockutils [req-6eb76848-47e7-4e7d-b7e3-4cb7b8f3b89e req-b49b125f-cfd0-491d-89f1-bcd8fc5dbc8c service nova] Releasing lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.732830] env[61728]: DEBUG nova.network.neutron [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1019.739933] env[61728]: DEBUG oslo_vmware.api [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464700, 'name': PowerOnVM_Task, 'duration_secs': 0.37138} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.740900] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1019.793617] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f381da29-d82f-4ea6-bf93-6bcf1d280b3d tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.794631] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.506s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.796865] env[61728]: INFO nova.compute.claims [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1019.801452] env[61728]: DEBUG nova.compute.manager [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1019.923685] env[61728]: DEBUG nova.network.neutron [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Updating instance_info_cache with network_info: [{"id": "1db8e5be-43b2-4fa8-954e-dce8e21ff0d4", "address": "fa:16:3e:25:be:c9", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1db8e5be-43", "ovs_interfaceid": "1db8e5be-43b2-4fa8-954e-dce8e21ff0d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.958952] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1019.959224] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1019.959485] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.959674] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.959875] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1019.960150] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7b56790-422f-4363-b78c-650e2580d603 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.969051] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1019.969051] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1019.969265] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7bfe421-0151-4198-9aac-9f689761a710 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.974919] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1019.974919] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5281f8dc-fe90-e188-8264-b83937086153" [ 1019.974919] env[61728]: _type = "Task" [ 1019.974919] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.983151] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5281f8dc-fe90-e188-8264-b83937086153, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.400579] env[61728]: DEBUG nova.network.neutron [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Successfully updated port: d0b341ea-4a0b-46ff-89f2-f96d212d04d9 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1020.426809] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "refresh_cache-bb7b38d5-53c7-4c05-96f8-9121171d3c46" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.427189] env[61728]: DEBUG nova.compute.manager [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Instance network_info: |[{"id": "1db8e5be-43b2-4fa8-954e-dce8e21ff0d4", "address": "fa:16:3e:25:be:c9", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1db8e5be-43", "ovs_interfaceid": "1db8e5be-43b2-4fa8-954e-dce8e21ff0d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1020.427648] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:be:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc6d5964-1106-4345-a26d-185dabd4ff0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1db8e5be-43b2-4fa8-954e-dce8e21ff0d4', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.435127] env[61728]: DEBUG oslo.service.loopingcall [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.435604] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1020.435889] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d657f0c2-0183-4d5b-9c4a-8924ef80c236 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.454951] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.454951] env[61728]: value = "task-464702" [ 1020.454951] env[61728]: _type = "Task" [ 1020.454951] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.462680] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464702, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.484610] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5281f8dc-fe90-e188-8264-b83937086153, 'name': SearchDatastore_Task, 'duration_secs': 0.011732} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.485605] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f65e46ac-9443-448f-820f-8b074112def1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.490861] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1020.490861] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528b14aa-adfc-024b-bfbf-afbd659a7599" [ 1020.490861] env[61728]: _type = "Task" [ 1020.490861] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.497886] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528b14aa-adfc-024b-bfbf-afbd659a7599, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.708341] env[61728]: DEBUG nova.compute.manager [req-dd430f3e-3088-496a-b5cd-1bd693128673 req-2c9a97bb-0b8f-4627-b24c-f21a3fa9bab2 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Received event network-vif-plugged-d0b341ea-4a0b-46ff-89f2-f96d212d04d9 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.708648] env[61728]: DEBUG oslo_concurrency.lockutils [req-dd430f3e-3088-496a-b5cd-1bd693128673 req-2c9a97bb-0b8f-4627-b24c-f21a3fa9bab2 service nova] Acquiring lock "705c8838-f605-46fa-b036-ec212f673f66-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.708965] env[61728]: DEBUG oslo_concurrency.lockutils [req-dd430f3e-3088-496a-b5cd-1bd693128673 req-2c9a97bb-0b8f-4627-b24c-f21a3fa9bab2 service nova] Lock "705c8838-f605-46fa-b036-ec212f673f66-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.709248] env[61728]: DEBUG oslo_concurrency.lockutils [req-dd430f3e-3088-496a-b5cd-1bd693128673 req-2c9a97bb-0b8f-4627-b24c-f21a3fa9bab2 service nova] Lock "705c8838-f605-46fa-b036-ec212f673f66-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.709517] env[61728]: DEBUG nova.compute.manager [req-dd430f3e-3088-496a-b5cd-1bd693128673 req-2c9a97bb-0b8f-4627-b24c-f21a3fa9bab2 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] No waiting events found dispatching network-vif-plugged-d0b341ea-4a0b-46ff-89f2-f96d212d04d9 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1020.709814] env[61728]: WARNING nova.compute.manager [req-dd430f3e-3088-496a-b5cd-1bd693128673 req-2c9a97bb-0b8f-4627-b24c-f21a3fa9bab2 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Received unexpected event network-vif-plugged-d0b341ea-4a0b-46ff-89f2-f96d212d04d9 for instance with vm_state active and task_state None. [ 1020.710055] env[61728]: DEBUG nova.compute.manager [req-dd430f3e-3088-496a-b5cd-1bd693128673 req-2c9a97bb-0b8f-4627-b24c-f21a3fa9bab2 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Received event network-changed-d0b341ea-4a0b-46ff-89f2-f96d212d04d9 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.710239] env[61728]: DEBUG nova.compute.manager [req-dd430f3e-3088-496a-b5cd-1bd693128673 req-2c9a97bb-0b8f-4627-b24c-f21a3fa9bab2 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Refreshing instance network info cache due to event network-changed-d0b341ea-4a0b-46ff-89f2-f96d212d04d9. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1020.710486] env[61728]: DEBUG oslo_concurrency.lockutils [req-dd430f3e-3088-496a-b5cd-1bd693128673 req-2c9a97bb-0b8f-4627-b24c-f21a3fa9bab2 service nova] Acquiring lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.710740] env[61728]: DEBUG oslo_concurrency.lockutils [req-dd430f3e-3088-496a-b5cd-1bd693128673 req-2c9a97bb-0b8f-4627-b24c-f21a3fa9bab2 service nova] Acquired lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.710993] env[61728]: DEBUG nova.network.neutron [req-dd430f3e-3088-496a-b5cd-1bd693128673 req-2c9a97bb-0b8f-4627-b24c-f21a3fa9bab2 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Refreshing network info cache for port d0b341ea-4a0b-46ff-89f2-f96d212d04d9 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1020.775152] env[61728]: DEBUG nova.compute.manager [req-5ee82f9e-04cc-4fe9-abdc-c1f651d49b80 req-c7298c7e-d091-4fc9-9f17-a209c354c561 service nova] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Received event network-changed-1db8e5be-43b2-4fa8-954e-dce8e21ff0d4 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.775417] env[61728]: DEBUG nova.compute.manager [req-5ee82f9e-04cc-4fe9-abdc-c1f651d49b80 req-c7298c7e-d091-4fc9-9f17-a209c354c561 service nova] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Refreshing instance network info cache due to event network-changed-1db8e5be-43b2-4fa8-954e-dce8e21ff0d4. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1020.775685] env[61728]: DEBUG oslo_concurrency.lockutils [req-5ee82f9e-04cc-4fe9-abdc-c1f651d49b80 req-c7298c7e-d091-4fc9-9f17-a209c354c561 service nova] Acquiring lock "refresh_cache-bb7b38d5-53c7-4c05-96f8-9121171d3c46" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.775879] env[61728]: DEBUG oslo_concurrency.lockutils [req-5ee82f9e-04cc-4fe9-abdc-c1f651d49b80 req-c7298c7e-d091-4fc9-9f17-a209c354c561 service nova] Acquired lock "refresh_cache-bb7b38d5-53c7-4c05-96f8-9121171d3c46" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.776100] env[61728]: DEBUG nova.network.neutron [req-5ee82f9e-04cc-4fe9-abdc-c1f651d49b80 req-c7298c7e-d091-4fc9-9f17-a209c354c561 service nova] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Refreshing network info cache for port 1db8e5be-43b2-4fa8-954e-dce8e21ff0d4 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1020.787368] env[61728]: INFO nova.compute.manager [None req-9370a620-124d-4c5a-9f7a-195df9f95cf5 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance to original state: 'active' [ 1020.809118] env[61728]: DEBUG nova.compute.manager [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1020.838071] env[61728]: DEBUG nova.virt.hardware [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1020.838454] env[61728]: DEBUG nova.virt.hardware [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1020.838595] env[61728]: DEBUG nova.virt.hardware [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.838833] env[61728]: DEBUG nova.virt.hardware [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1020.839038] env[61728]: DEBUG nova.virt.hardware [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.839246] env[61728]: DEBUG nova.virt.hardware [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1020.839508] env[61728]: DEBUG nova.virt.hardware [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1020.839722] env[61728]: DEBUG nova.virt.hardware [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1020.839926] env[61728]: DEBUG nova.virt.hardware [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1020.840149] env[61728]: DEBUG nova.virt.hardware [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1020.840372] env[61728]: DEBUG nova.virt.hardware [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.841365] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588c3164-7e59-4dca-a69c-1f4905442ada {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.853716] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba46c3f-a406-4e68-8e21-f5751b5c311d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.903607] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.964553] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464702, 'name': CreateVM_Task, 'duration_secs': 0.389025} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.966861] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1020.968835] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.970555] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.970555] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1020.970555] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9da3a3d9-9bde-4af5-a2d4-673d059455bb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.974996] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1020.974996] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520d8a74-c87e-94fd-229b-1d0f7db3c0de" [ 1020.974996] env[61728]: _type = "Task" [ 1020.974996] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.985925] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520d8a74-c87e-94fd-229b-1d0f7db3c0de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.000603] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528b14aa-adfc-024b-bfbf-afbd659a7599, 'name': SearchDatastore_Task, 'duration_secs': 0.008723} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.003304] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.003580] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk. {{(pid=61728) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1021.004128] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c3d49cd3-4a27-4067-b8a5-5dbd2c2107f7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.011533] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1021.011533] env[61728]: value = "task-464703" [ 1021.011533] env[61728]: _type = "Task" [ 1021.011533] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.021373] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464703, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.038797] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f369fa5-47b8-48cc-b54e-df535033d01d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.046575] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-690860b4-36bb-4b54-a96f-673de94e4556 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.079660] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f2adc3-de2c-4008-86f2-f79402060f7d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.096099] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d821af90-b82c-4672-a922-7af745005fcf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.111131] env[61728]: DEBUG nova.compute.provider_tree [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1021.166801] env[61728]: DEBUG nova.network.neutron [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Successfully updated port: 216193d1-1332-49ba-8702-dacc02486ec1 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1021.446470] env[61728]: DEBUG nova.network.neutron [req-dd430f3e-3088-496a-b5cd-1bd693128673 req-2c9a97bb-0b8f-4627-b24c-f21a3fa9bab2 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Added VIF to instance network info cache for port d0b341ea-4a0b-46ff-89f2-f96d212d04d9. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3489}} [ 1021.446470] env[61728]: DEBUG nova.network.neutron [req-dd430f3e-3088-496a-b5cd-1bd693128673 req-2c9a97bb-0b8f-4627-b24c-f21a3fa9bab2 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Updating instance_info_cache with network_info: [{"id": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "address": "fa:16:3e:57:58:2b", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec0ae08d-62", "ovs_interfaceid": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d0b341ea-4a0b-46ff-89f2-f96d212d04d9", "address": "fa:16:3e:0a:de:a7", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0b341ea-4a", "ovs_interfaceid": "d0b341ea-4a0b-46ff-89f2-f96d212d04d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.488261] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520d8a74-c87e-94fd-229b-1d0f7db3c0de, 'name': SearchDatastore_Task, 'duration_secs': 0.009868} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.488590] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.488826] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.489106] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.489283] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.489473] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.489765] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0bb0566-894c-4bcc-badf-3c541da85510 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.505781] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.505781] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1021.506391] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fdf773a-7ab1-43d6-9e1e-23152b580d63 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.511830] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1021.511830] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b711a5-e17d-159e-33cc-c752cbb5779e" [ 1021.511830] env[61728]: _type = "Task" [ 1021.511830] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.522306] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b711a5-e17d-159e-33cc-c752cbb5779e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.525864] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464703, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.617575] env[61728]: DEBUG nova.network.neutron [req-5ee82f9e-04cc-4fe9-abdc-c1f651d49b80 req-c7298c7e-d091-4fc9-9f17-a209c354c561 service nova] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Updated VIF entry in instance network info cache for port 1db8e5be-43b2-4fa8-954e-dce8e21ff0d4. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1021.617944] env[61728]: DEBUG nova.network.neutron [req-5ee82f9e-04cc-4fe9-abdc-c1f651d49b80 req-c7298c7e-d091-4fc9-9f17-a209c354c561 service nova] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Updating instance_info_cache with network_info: [{"id": "1db8e5be-43b2-4fa8-954e-dce8e21ff0d4", "address": "fa:16:3e:25:be:c9", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1db8e5be-43", "ovs_interfaceid": "1db8e5be-43b2-4fa8-954e-dce8e21ff0d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.636083] env[61728]: ERROR nova.scheduler.client.report [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [req-06a5931c-d459-4ce8-be3c-2ffc9dffe164] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-06a5931c-d459-4ce8-be3c-2ffc9dffe164"}]} [ 1021.651553] env[61728]: DEBUG nova.scheduler.client.report [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1021.665691] env[61728]: DEBUG nova.scheduler.client.report [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1021.665949] env[61728]: DEBUG nova.compute.provider_tree [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 113, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1021.670387] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Acquiring lock "refresh_cache-9efdee89-4c88-4af3-af04-5281e8f6bc25" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.670523] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Acquired lock "refresh_cache-9efdee89-4c88-4af3-af04-5281e8f6bc25" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.670668] env[61728]: DEBUG nova.network.neutron [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1021.678057] env[61728]: DEBUG nova.scheduler.client.report [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1021.697012] env[61728]: DEBUG nova.scheduler.client.report [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1021.905564] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ca8920-ef66-430f-8225-acd9027149a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.913196] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b46224f-7813-474c-95b4-1166ee87b2e6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.942818] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5222c70b-4205-48a5-979f-1d6ac9b56ded {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.949952] env[61728]: DEBUG oslo_concurrency.lockutils [req-dd430f3e-3088-496a-b5cd-1bd693128673 req-2c9a97bb-0b8f-4627-b24c-f21a3fa9bab2 service nova] Releasing lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.950467] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.950666] env[61728]: DEBUG nova.network.neutron [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1021.952823] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5f7c0f-3670-469c-b871-84dbb43468fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.967484] env[61728]: DEBUG nova.compute.provider_tree [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1022.024557] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52b711a5-e17d-159e-33cc-c752cbb5779e, 'name': SearchDatastore_Task, 'duration_secs': 0.032902} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.028124] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464703, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519634} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.028348] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59f3849f-ce91-41cc-8db5-9e35321c666c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.030418] env[61728]: INFO nova.virt.vmwareapi.ds_util [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk. [ 1022.031141] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36bed4af-5ed9-4227-bac8-c1ef971e1ee1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.036791] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1022.036791] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52cc52fa-f26e-6253-e91b-a8075bb8db88" [ 1022.036791] env[61728]: _type = "Task" [ 1022.036791] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.058976] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.063577] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90d47a94-bee0-4d19-a046-b316d6172f9e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.084354] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52cc52fa-f26e-6253-e91b-a8075bb8db88, 'name': SearchDatastore_Task, 'duration_secs': 0.010138} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.084354] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.084596] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] bb7b38d5-53c7-4c05-96f8-9121171d3c46/bb7b38d5-53c7-4c05-96f8-9121171d3c46.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1022.084885] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1022.084885] env[61728]: value = "task-464704" [ 1022.084885] env[61728]: _type = "Task" [ 1022.084885] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.085093] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-144cd031-15a2-4296-8357-c28718e79606 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.095754] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464704, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.096788] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1022.096788] env[61728]: value = "task-464705" [ 1022.096788] env[61728]: _type = "Task" [ 1022.096788] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.104243] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464705, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.120649] env[61728]: DEBUG oslo_concurrency.lockutils [req-5ee82f9e-04cc-4fe9-abdc-c1f651d49b80 req-c7298c7e-d091-4fc9-9f17-a209c354c561 service nova] Releasing lock "refresh_cache-bb7b38d5-53c7-4c05-96f8-9121171d3c46" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.202618] env[61728]: DEBUG nova.network.neutron [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1022.375490] env[61728]: DEBUG nova.network.neutron [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Updating instance_info_cache with network_info: [{"id": "216193d1-1332-49ba-8702-dacc02486ec1", "address": "fa:16:3e:4c:d3:d7", "network": {"id": "41f527d6-7a12-4374-85e7-94194af385b5", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-260792778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9f255ab8e7a40cb86b7bdb1f3aadf2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53955f0e-c162-4cef-8bd5-335b369c36b6", "external-id": "nsx-vlan-transportzone-623", "segmentation_id": 623, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap216193d1-13", "ovs_interfaceid": "216193d1-1332-49ba-8702-dacc02486ec1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.488788] env[61728]: WARNING nova.network.neutron [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] 20ba9b61-3be9-4ad2-a1d0-88c810873bcb already exists in list: networks containing: ['20ba9b61-3be9-4ad2-a1d0-88c810873bcb']. ignoring it [ 1022.489052] env[61728]: WARNING nova.network.neutron [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] 20ba9b61-3be9-4ad2-a1d0-88c810873bcb already exists in list: networks containing: ['20ba9b61-3be9-4ad2-a1d0-88c810873bcb']. ignoring it [ 1022.489249] env[61728]: WARNING nova.network.neutron [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] d0b341ea-4a0b-46ff-89f2-f96d212d04d9 already exists in list: port_ids containing: ['d0b341ea-4a0b-46ff-89f2-f96d212d04d9']. ignoring it [ 1022.498763] env[61728]: DEBUG nova.scheduler.client.report [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 160 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1022.499086] env[61728]: DEBUG nova.compute.provider_tree [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 160 to 161 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1022.499297] env[61728]: DEBUG nova.compute.provider_tree [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1022.600153] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464704, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.611354] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464705, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.879954] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Releasing lock "refresh_cache-9efdee89-4c88-4af3-af04-5281e8f6bc25" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.880310] env[61728]: DEBUG nova.compute.manager [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Instance network_info: |[{"id": "216193d1-1332-49ba-8702-dacc02486ec1", "address": "fa:16:3e:4c:d3:d7", "network": {"id": "41f527d6-7a12-4374-85e7-94194af385b5", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-260792778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9f255ab8e7a40cb86b7bdb1f3aadf2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53955f0e-c162-4cef-8bd5-335b369c36b6", "external-id": "nsx-vlan-transportzone-623", "segmentation_id": 623, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap216193d1-13", "ovs_interfaceid": "216193d1-1332-49ba-8702-dacc02486ec1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1022.880742] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4c:d3:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '53955f0e-c162-4cef-8bd5-335b369c36b6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '216193d1-1332-49ba-8702-dacc02486ec1', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1022.888394] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Creating folder: Project (c9f255ab8e7a40cb86b7bdb1f3aadf2c). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1022.888711] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d5b37931-ae09-4ed0-8327-2eff19da039b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.900663] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Created folder: Project (c9f255ab8e7a40cb86b7bdb1f3aadf2c) in parent group-v121913. [ 1022.900914] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Creating folder: Instances. Parent ref: group-v122184. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1022.901221] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-622709c2-5abd-4c8f-a1ed-0a5869a49905 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.909987] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Created folder: Instances in parent group-v122184. [ 1022.910254] env[61728]: DEBUG oslo.service.loopingcall [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.910432] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1022.910635] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4afdedf-6ee6-4957-a38a-02c54b6c114c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.931645] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1022.931645] env[61728]: value = "task-464708" [ 1022.931645] env[61728]: _type = "Task" [ 1022.931645] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.939467] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464708, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.970740] env[61728]: DEBUG nova.network.neutron [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Updating instance_info_cache with network_info: [{"id": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "address": "fa:16:3e:57:58:2b", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec0ae08d-62", "ovs_interfaceid": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d0b341ea-4a0b-46ff-89f2-f96d212d04d9", "address": "fa:16:3e:0a:de:a7", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0b341ea-4a", "ovs_interfaceid": "d0b341ea-4a0b-46ff-89f2-f96d212d04d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.987295] env[61728]: DEBUG nova.compute.manager [req-c6e5d144-6a37-4341-a1d3-ce62b9f29eb3 req-e5926c6c-a184-4a26-8841-5c68d47a84f0 service nova] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Received event network-vif-plugged-216193d1-1332-49ba-8702-dacc02486ec1 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.987499] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6e5d144-6a37-4341-a1d3-ce62b9f29eb3 req-e5926c6c-a184-4a26-8841-5c68d47a84f0 service nova] Acquiring lock "9efdee89-4c88-4af3-af04-5281e8f6bc25-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.987672] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6e5d144-6a37-4341-a1d3-ce62b9f29eb3 req-e5926c6c-a184-4a26-8841-5c68d47a84f0 service nova] Lock "9efdee89-4c88-4af3-af04-5281e8f6bc25-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.987981] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6e5d144-6a37-4341-a1d3-ce62b9f29eb3 req-e5926c6c-a184-4a26-8841-5c68d47a84f0 service nova] Lock "9efdee89-4c88-4af3-af04-5281e8f6bc25-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.988337] env[61728]: DEBUG nova.compute.manager [req-c6e5d144-6a37-4341-a1d3-ce62b9f29eb3 req-e5926c6c-a184-4a26-8841-5c68d47a84f0 service nova] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] No waiting events found dispatching network-vif-plugged-216193d1-1332-49ba-8702-dacc02486ec1 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1022.988539] env[61728]: WARNING nova.compute.manager [req-c6e5d144-6a37-4341-a1d3-ce62b9f29eb3 req-e5926c6c-a184-4a26-8841-5c68d47a84f0 service nova] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Received unexpected event network-vif-plugged-216193d1-1332-49ba-8702-dacc02486ec1 for instance with vm_state building and task_state spawning. [ 1022.988779] env[61728]: DEBUG nova.compute.manager [req-c6e5d144-6a37-4341-a1d3-ce62b9f29eb3 req-e5926c6c-a184-4a26-8841-5c68d47a84f0 service nova] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Received event network-changed-216193d1-1332-49ba-8702-dacc02486ec1 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.988987] env[61728]: DEBUG nova.compute.manager [req-c6e5d144-6a37-4341-a1d3-ce62b9f29eb3 req-e5926c6c-a184-4a26-8841-5c68d47a84f0 service nova] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Refreshing instance network info cache due to event network-changed-216193d1-1332-49ba-8702-dacc02486ec1. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1022.989221] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6e5d144-6a37-4341-a1d3-ce62b9f29eb3 req-e5926c6c-a184-4a26-8841-5c68d47a84f0 service nova] Acquiring lock "refresh_cache-9efdee89-4c88-4af3-af04-5281e8f6bc25" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.989368] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6e5d144-6a37-4341-a1d3-ce62b9f29eb3 req-e5926c6c-a184-4a26-8841-5c68d47a84f0 service nova] Acquired lock "refresh_cache-9efdee89-4c88-4af3-af04-5281e8f6bc25" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.989581] env[61728]: DEBUG nova.network.neutron [req-c6e5d144-6a37-4341-a1d3-ce62b9f29eb3 req-e5926c6c-a184-4a26-8841-5c68d47a84f0 service nova] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Refreshing network info cache for port 216193d1-1332-49ba-8702-dacc02486ec1 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1023.005540] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.210s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.005540] env[61728]: DEBUG nova.compute.manager [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1023.008777] env[61728]: DEBUG oslo_concurrency.lockutils [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.150s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.009020] env[61728]: DEBUG nova.objects.instance [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lazy-loading 'resources' on Instance uuid 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.098907] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464704, 'name': ReconfigVM_Task, 'duration_secs': 0.97304} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.099094] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1023.099958] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181e1408-c3a0-44d9-a589-19c31fcc38d4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.110290] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464705, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.684029} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.128364] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] bb7b38d5-53c7-4c05-96f8-9121171d3c46/bb7b38d5-53c7-4c05-96f8-9121171d3c46.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1023.128595] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1023.133724] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0b632dea-3a47-4016-9c77-efd6f5b17369 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.135558] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c301bd9b-efe8-4307-99ec-7c21fa9ce107 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.150609] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1023.150609] env[61728]: value = "task-464710" [ 1023.150609] env[61728]: _type = "Task" [ 1023.150609] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.151954] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1023.151954] env[61728]: value = "task-464709" [ 1023.151954] env[61728]: _type = "Task" [ 1023.151954] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.162797] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464710, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.165639] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464709, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.284570] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "f9f07464-9f43-43fd-8895-2673861747fa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.284839] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "f9f07464-9f43-43fd-8895-2673861747fa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.285079] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "f9f07464-9f43-43fd-8895-2673861747fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.285277] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "f9f07464-9f43-43fd-8895-2673861747fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.285455] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "f9f07464-9f43-43fd-8895-2673861747fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.287731] env[61728]: INFO nova.compute.manager [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Terminating instance [ 1023.289680] env[61728]: DEBUG nova.compute.manager [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1023.289921] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1023.290993] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41fea26a-5447-4dfc-9405-8e4263383217 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.297724] env[61728]: DEBUG oslo_vmware.api [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1023.297724] env[61728]: value = "task-464711" [ 1023.297724] env[61728]: _type = "Task" [ 1023.297724] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.306498] env[61728]: DEBUG oslo_vmware.api [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464711, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.441622] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464708, 'name': CreateVM_Task, 'duration_secs': 0.34416} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.441819] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1023.442544] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.442716] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.443058] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1023.443313] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6de78d6-6613-47d3-a26c-9a98f278b8ea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.448141] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Waiting for the task: (returnval){ [ 1023.448141] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52029ca1-8bb1-0127-077a-ffb697656952" [ 1023.448141] env[61728]: _type = "Task" [ 1023.448141] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.455688] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52029ca1-8bb1-0127-077a-ffb697656952, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.473424] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.474026] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.474193] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.474941] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-977fb033-3cf8-4fbd-9d68-d98eb45ddd3c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.490431] env[61728]: DEBUG nova.virt.hardware [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1023.490672] env[61728]: DEBUG nova.virt.hardware [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1023.490836] env[61728]: DEBUG nova.virt.hardware [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1023.491058] env[61728]: DEBUG nova.virt.hardware [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1023.491229] env[61728]: DEBUG nova.virt.hardware [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1023.491394] env[61728]: DEBUG nova.virt.hardware [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1023.491604] env[61728]: DEBUG nova.virt.hardware [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1023.491813] env[61728]: DEBUG nova.virt.hardware [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1023.492008] env[61728]: DEBUG nova.virt.hardware [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1023.492187] env[61728]: DEBUG nova.virt.hardware [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1023.492370] env[61728]: DEBUG nova.virt.hardware [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1023.498540] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Reconfiguring VM to attach interface {{(pid=61728) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1023.500856] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8aed1ca-a1df-4fc2-b58f-60950ec6203d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.513507] env[61728]: DEBUG nova.compute.utils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1023.518424] env[61728]: DEBUG nova.compute.manager [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1023.518424] env[61728]: DEBUG nova.network.neutron [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1023.524829] env[61728]: DEBUG oslo_vmware.api [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 1023.524829] env[61728]: value = "task-464712" [ 1023.524829] env[61728]: _type = "Task" [ 1023.524829] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.536023] env[61728]: DEBUG oslo_vmware.api [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464712, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.567727] env[61728]: DEBUG nova.policy [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed72340258144426965685530073676c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4cbef4831f2c4e27bf644608817217b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1023.670208] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464709, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08094} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.673179] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1023.673540] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464710, 'name': ReconfigVM_Task, 'duration_secs': 0.225236} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.674324] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01a1bd6-c348-4aa9-bc28-21a46eec1f08 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.676896] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1023.681370] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-450028fb-232a-4e2a-a2da-cb7c4cbaeb8a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.707124] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] bb7b38d5-53c7-4c05-96f8-9121171d3c46/bb7b38d5-53c7-4c05-96f8-9121171d3c46.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.711104] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55c376e9-7e98-4d0f-83f6-80797f3ca20b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.728029] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1023.728029] env[61728]: value = "task-464713" [ 1023.728029] env[61728]: _type = "Task" [ 1023.728029] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.736752] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1023.736752] env[61728]: value = "task-464714" [ 1023.736752] env[61728]: _type = "Task" [ 1023.736752] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.742070] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464713, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.747967] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464714, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.805375] env[61728]: DEBUG nova.network.neutron [req-c6e5d144-6a37-4341-a1d3-ce62b9f29eb3 req-e5926c6c-a184-4a26-8841-5c68d47a84f0 service nova] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Updated VIF entry in instance network info cache for port 216193d1-1332-49ba-8702-dacc02486ec1. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1023.805766] env[61728]: DEBUG nova.network.neutron [req-c6e5d144-6a37-4341-a1d3-ce62b9f29eb3 req-e5926c6c-a184-4a26-8841-5c68d47a84f0 service nova] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Updating instance_info_cache with network_info: [{"id": "216193d1-1332-49ba-8702-dacc02486ec1", "address": "fa:16:3e:4c:d3:d7", "network": {"id": "41f527d6-7a12-4374-85e7-94194af385b5", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-260792778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9f255ab8e7a40cb86b7bdb1f3aadf2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53955f0e-c162-4cef-8bd5-335b369c36b6", "external-id": "nsx-vlan-transportzone-623", "segmentation_id": 623, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap216193d1-13", "ovs_interfaceid": "216193d1-1332-49ba-8702-dacc02486ec1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.812713] env[61728]: DEBUG oslo_vmware.api [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464711, 'name': PowerOffVM_Task, 'duration_secs': 0.165611} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.812980] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1023.813208] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1023.813476] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122162', 'volume_id': 'fe562e41-06ae-47c9-9eda-4f075c89cd34', 'name': 'volume-fe562e41-06ae-47c9-9eda-4f075c89cd34', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'f9f07464-9f43-43fd-8895-2673861747fa', 'attached_at': '2024-08-29T12:31:14.000000', 'detached_at': '', 'volume_id': 'fe562e41-06ae-47c9-9eda-4f075c89cd34', 'serial': 'fe562e41-06ae-47c9-9eda-4f075c89cd34'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1023.815215] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ecb16a-6da4-4373-9975-f742658b021b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.838531] env[61728]: DEBUG nova.network.neutron [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Successfully created port: adc8e830-5757-48a7-8060-af7d189243df {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1023.840880] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6020648-de45-41ab-b322-22170f5149d4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.849672] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3d7250-f65a-48a2-9840-0d0a6b4e3494 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.871941] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0728a3a-a6e0-4231-99e3-e4c350cbd8a9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.875223] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3479f76-a57c-47ef-b62f-24c34f451045 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.894366] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb2880f-3a71-49e3-b275-c628b9361f86 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.898226] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] The volume has not been displaced from its original location: [datastore1] volume-fe562e41-06ae-47c9-9eda-4f075c89cd34/volume-fe562e41-06ae-47c9-9eda-4f075c89cd34.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1023.904114] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Reconfiguring VM instance instance-00000051 to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1023.904503] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07fea831-9017-4251-b339-012b7ff9c8bd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.949581] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91eb93b7-a9e7-48b3-b975-0479fb13ddce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.952681] env[61728]: DEBUG oslo_vmware.api [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1023.952681] env[61728]: value = "task-464715" [ 1023.952681] env[61728]: _type = "Task" [ 1023.952681] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.962663] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07966ce6-1bfd-40b8-b217-fed8202104d0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.973201] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52029ca1-8bb1-0127-077a-ffb697656952, 'name': SearchDatastore_Task, 'duration_secs': 0.021278} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.973558] env[61728]: DEBUG oslo_vmware.api [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464715, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.975460] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.975460] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1023.975460] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.975460] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.975677] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1023.983891] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fd0f3cfc-31fd-4e25-a1b4-ac64d39eacb1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.986387] env[61728]: DEBUG nova.compute.provider_tree [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.994709] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1023.994921] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1023.995779] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-392ff614-6206-4627-822e-de0b7513399f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.002572] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Waiting for the task: (returnval){ [ 1024.002572] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f4c2f8-ad86-f1f0-b75c-ff369748e59d" [ 1024.002572] env[61728]: _type = "Task" [ 1024.002572] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.012526] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f4c2f8-ad86-f1f0-b75c-ff369748e59d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.018379] env[61728]: DEBUG nova.compute.manager [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1024.036861] env[61728]: DEBUG oslo_vmware.api [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464712, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.248497] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464714, 'name': ReconfigVM_Task, 'duration_secs': 0.426017} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.249254] env[61728]: DEBUG oslo_vmware.api [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464713, 'name': PowerOnVM_Task, 'duration_secs': 0.466079} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.249353] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Reconfigured VM instance instance-00000066 to attach disk [datastore1] bb7b38d5-53c7-4c05-96f8-9121171d3c46/bb7b38d5-53c7-4c05-96f8-9121171d3c46.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.251023] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1024.251895] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee7e748e-c19c-4f92-931d-a0c18f0b470d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.254113] env[61728]: DEBUG nova.compute.manager [None req-89fbe982-6386-415a-8eb3-a87e6d9cb130 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.254846] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbef46f6-ae65-454e-9bc0-526c77845803 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.264275] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1024.264275] env[61728]: value = "task-464716" [ 1024.264275] env[61728]: _type = "Task" [ 1024.264275] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.271863] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464716, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.308564] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6e5d144-6a37-4341-a1d3-ce62b9f29eb3 req-e5926c6c-a184-4a26-8841-5c68d47a84f0 service nova] Releasing lock "refresh_cache-9efdee89-4c88-4af3-af04-5281e8f6bc25" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.466683] env[61728]: DEBUG oslo_vmware.api [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464715, 'name': ReconfigVM_Task, 'duration_secs': 0.279135} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.466999] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Reconfigured VM instance instance-00000051 to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1024.471751] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18a81ccc-50ca-4709-a21b-30446dbec07b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.486575] env[61728]: DEBUG oslo_vmware.api [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1024.486575] env[61728]: value = "task-464717" [ 1024.486575] env[61728]: _type = "Task" [ 1024.486575] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.490109] env[61728]: DEBUG nova.scheduler.client.report [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.499508] env[61728]: DEBUG oslo_vmware.api [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464717, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.514685] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f4c2f8-ad86-f1f0-b75c-ff369748e59d, 'name': SearchDatastore_Task, 'duration_secs': 0.010486} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.515660] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5360cfdd-dd89-48b9-bf21-bebf54fa791a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.521650] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Waiting for the task: (returnval){ [ 1024.521650] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5278d3c2-0805-bbd6-87cc-03003c6f3011" [ 1024.521650] env[61728]: _type = "Task" [ 1024.521650] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.539632] env[61728]: DEBUG oslo_vmware.api [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464712, 'name': ReconfigVM_Task, 'duration_secs': 0.5598} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.539958] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5278d3c2-0805-bbd6-87cc-03003c6f3011, 'name': SearchDatastore_Task, 'duration_secs': 0.013647} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.540450] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.540663] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Reconfigured VM to attach interface {{(pid=61728) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1024.543147] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.543576] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 9efdee89-4c88-4af3-af04-5281e8f6bc25/9efdee89-4c88-4af3-af04-5281e8f6bc25.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1024.543841] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1b3a4e9-4eb2-4ef2-89da-9a2e6ce735fc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.551570] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Waiting for the task: (returnval){ [ 1024.551570] env[61728]: value = "task-464718" [ 1024.551570] env[61728]: _type = "Task" [ 1024.551570] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.559258] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464718, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.778205] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464716, 'name': Rename_Task, 'duration_secs': 0.143349} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.778497] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1024.778744] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-28608b1d-b829-4560-8760-41da6940ca70 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.785310] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1024.785310] env[61728]: value = "task-464719" [ 1024.785310] env[61728]: _type = "Task" [ 1024.785310] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.794189] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464719, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.998344] env[61728]: DEBUG oslo_concurrency.lockutils [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.989s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.000722] env[61728]: DEBUG oslo_vmware.api [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464717, 'name': ReconfigVM_Task, 'duration_secs': 0.163631} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.001311] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.453s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.001603] env[61728]: DEBUG nova.objects.instance [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lazy-loading 'resources' on Instance uuid 378ef752-6d47-4861-baf1-064b2a180dcc {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.002695] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122162', 'volume_id': 'fe562e41-06ae-47c9-9eda-4f075c89cd34', 'name': 'volume-fe562e41-06ae-47c9-9eda-4f075c89cd34', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'f9f07464-9f43-43fd-8895-2673861747fa', 'attached_at': '2024-08-29T12:31:14.000000', 'detached_at': '', 'volume_id': 'fe562e41-06ae-47c9-9eda-4f075c89cd34', 'serial': 'fe562e41-06ae-47c9-9eda-4f075c89cd34'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1025.003054] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1025.004158] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8350c8-311d-441a-af2d-0fdca94ad9b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.011658] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1025.012615] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07faf8ec-e999-4e1c-bdb8-c6ea68277e75 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.024394] env[61728]: INFO nova.scheduler.client.report [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleted allocations for instance 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f [ 1025.031299] env[61728]: DEBUG nova.compute.manager [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1025.046167] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d7735516-74e5-450a-9a39-db8528478ff8 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-705c8838-f605-46fa-b036-ec212f673f66-d0b341ea-4a0b-46ff-89f2-f96d212d04d9" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.060s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.065970] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464718, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.068635] env[61728]: DEBUG nova.virt.hardware [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1025.068970] env[61728]: DEBUG nova.virt.hardware [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1025.069157] env[61728]: DEBUG nova.virt.hardware [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1025.069399] env[61728]: DEBUG nova.virt.hardware [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1025.070260] env[61728]: DEBUG nova.virt.hardware [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1025.070260] env[61728]: DEBUG nova.virt.hardware [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1025.070260] env[61728]: DEBUG nova.virt.hardware [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1025.070260] env[61728]: DEBUG nova.virt.hardware [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1025.070456] env[61728]: DEBUG nova.virt.hardware [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1025.070617] env[61728]: DEBUG nova.virt.hardware [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1025.070863] env[61728]: DEBUG nova.virt.hardware [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1025.072500] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399fe76e-8806-4dde-a50c-1e6014782fb2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.083083] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca89e87-4f53-4a59-a542-05592c160f38 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.107157] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1025.107595] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1025.107921] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleting the datastore file [datastore1] f9f07464-9f43-43fd-8895-2673861747fa {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.108613] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49e91bac-4d8e-4899-98ff-96035d52fba3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.115886] env[61728]: DEBUG oslo_vmware.api [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1025.115886] env[61728]: value = "task-464721" [ 1025.115886] env[61728]: _type = "Task" [ 1025.115886] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.126322] env[61728]: DEBUG oslo_vmware.api [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464721, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.295780] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464719, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.354949] env[61728]: DEBUG nova.network.neutron [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Successfully updated port: adc8e830-5757-48a7-8060-af7d189243df {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1025.358559] env[61728]: DEBUG nova.compute.manager [req-c6b5fcc5-96b6-417d-9f42-a4f6360d5688 req-1a3e3dc2-69d7-4460-a499-77c45f248b54 service nova] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Received event network-vif-plugged-adc8e830-5757-48a7-8060-af7d189243df {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1025.358785] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6b5fcc5-96b6-417d-9f42-a4f6360d5688 req-1a3e3dc2-69d7-4460-a499-77c45f248b54 service nova] Acquiring lock "5f248cf6-31a9-49ef-a8a8-218987664af7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.359060] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6b5fcc5-96b6-417d-9f42-a4f6360d5688 req-1a3e3dc2-69d7-4460-a499-77c45f248b54 service nova] Lock "5f248cf6-31a9-49ef-a8a8-218987664af7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.359255] env[61728]: DEBUG oslo_concurrency.lockutils [req-c6b5fcc5-96b6-417d-9f42-a4f6360d5688 req-1a3e3dc2-69d7-4460-a499-77c45f248b54 service nova] Lock "5f248cf6-31a9-49ef-a8a8-218987664af7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.359402] env[61728]: DEBUG nova.compute.manager [req-c6b5fcc5-96b6-417d-9f42-a4f6360d5688 req-1a3e3dc2-69d7-4460-a499-77c45f248b54 service nova] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] No waiting events found dispatching network-vif-plugged-adc8e830-5757-48a7-8060-af7d189243df {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1025.359642] env[61728]: WARNING nova.compute.manager [req-c6b5fcc5-96b6-417d-9f42-a4f6360d5688 req-1a3e3dc2-69d7-4460-a499-77c45f248b54 service nova] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Received unexpected event network-vif-plugged-adc8e830-5757-48a7-8060-af7d189243df for instance with vm_state building and task_state spawning. [ 1025.539231] env[61728]: DEBUG oslo_concurrency.lockutils [None req-578924da-ef23-4138-a4f3-99ff6ce30142 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "0edd1f56-3e8e-4868-9c8e-a8caaf026e7f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.176s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.562484] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464718, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.685519} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.565122] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 9efdee89-4c88-4af3-af04-5281e8f6bc25/9efdee89-4c88-4af3-af04-5281e8f6bc25.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1025.565358] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1025.565859] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f18e5422-799f-4688-a542-b2d2e8ff9ffb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.572214] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Waiting for the task: (returnval){ [ 1025.572214] env[61728]: value = "task-464722" [ 1025.572214] env[61728]: _type = "Task" [ 1025.572214] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.579744] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464722, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.626676] env[61728]: DEBUG oslo_vmware.api [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464721, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.723630] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407606c2-911e-43a4-8c8d-373af466aa8a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.730827] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f5c41ef-3288-4a2d-afa7-9209cf884479 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.761972] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd21e67f-6ce6-4a48-9aae-49634eafa03f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.768991] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf04112-2d04-4fe6-9453-32b49fbfe67f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.782280] env[61728]: DEBUG nova.compute.provider_tree [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.785198] env[61728]: INFO nova.compute.manager [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Unrescuing [ 1025.785444] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.785602] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquired lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.785774] env[61728]: DEBUG nova.network.neutron [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1025.794808] env[61728]: DEBUG oslo_vmware.api [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464719, 'name': PowerOnVM_Task, 'duration_secs': 0.627981} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.795090] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1025.795272] env[61728]: INFO nova.compute.manager [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Took 7.39 seconds to spawn the instance on the hypervisor. [ 1025.795457] env[61728]: DEBUG nova.compute.manager [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1025.796408] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cdc6aea-65b8-4c7d-bee5-dba39a2c78b3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.857795] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Acquiring lock "refresh_cache-5f248cf6-31a9-49ef-a8a8-218987664af7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.857966] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Acquired lock "refresh_cache-5f248cf6-31a9-49ef-a8a8-218987664af7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.858122] env[61728]: DEBUG nova.network.neutron [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1026.084081] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464722, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06374} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.084386] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1026.085176] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a3417a-00d0-4e76-8381-262230fe45ec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.107041] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 9efdee89-4c88-4af3-af04-5281e8f6bc25/9efdee89-4c88-4af3-af04-5281e8f6bc25.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1026.107353] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b69ecf45-73f9-451d-89d9-dbd447d30e48 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.132503] env[61728]: DEBUG oslo_vmware.api [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464721, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.569418} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.134065] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.134280] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1026.134463] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1026.134662] env[61728]: INFO nova.compute.manager [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Took 2.84 seconds to destroy the instance on the hypervisor. [ 1026.134908] env[61728]: DEBUG oslo.service.loopingcall [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.135230] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Waiting for the task: (returnval){ [ 1026.135230] env[61728]: value = "task-464723" [ 1026.135230] env[61728]: _type = "Task" [ 1026.135230] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.135372] env[61728]: DEBUG nova.compute.manager [-] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1026.135459] env[61728]: DEBUG nova.network.neutron [-] [instance: f9f07464-9f43-43fd-8895-2673861747fa] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1026.145573] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464723, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.288062] env[61728]: DEBUG nova.scheduler.client.report [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.321761] env[61728]: INFO nova.compute.manager [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Took 22.59 seconds to build instance. [ 1026.404166] env[61728]: DEBUG nova.network.neutron [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1026.648590] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464723, 'name': ReconfigVM_Task, 'duration_secs': 0.269013} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.648911] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 9efdee89-4c88-4af3-af04-5281e8f6bc25/9efdee89-4c88-4af3-af04-5281e8f6bc25.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.649668] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5bd9f6df-01a4-4795-9034-e98922ba29be {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.656494] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Waiting for the task: (returnval){ [ 1026.656494] env[61728]: value = "task-464724" [ 1026.656494] env[61728]: _type = "Task" [ 1026.656494] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.666554] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464724, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.680348] env[61728]: DEBUG oslo_concurrency.lockutils [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.681533] env[61728]: DEBUG oslo_concurrency.lockutils [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.681981] env[61728]: INFO nova.compute.manager [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Rebooting instance [ 1026.763390] env[61728]: DEBUG nova.network.neutron [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Updating instance_info_cache with network_info: [{"id": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "address": "fa:16:3e:d1:31:88", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67077c45-d8", "ovs_interfaceid": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.793555] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.792s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.797066] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.436s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.798668] env[61728]: INFO nova.compute.claims [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1026.823353] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e5c40bd2-9d78-4c07-9741-44e478b4e326 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "bb7b38d5-53c7-4c05-96f8-9121171d3c46" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.105s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.832590] env[61728]: INFO nova.scheduler.client.report [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Deleted allocations for instance 378ef752-6d47-4861-baf1-064b2a180dcc [ 1026.848210] env[61728]: DEBUG oslo_concurrency.lockutils [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "interface-705c8838-f605-46fa-b036-ec212f673f66-d0b341ea-4a0b-46ff-89f2-f96d212d04d9" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.848467] env[61728]: DEBUG oslo_concurrency.lockutils [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-705c8838-f605-46fa-b036-ec212f673f66-d0b341ea-4a0b-46ff-89f2-f96d212d04d9" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.856713] env[61728]: DEBUG nova.network.neutron [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Updating instance_info_cache with network_info: [{"id": "adc8e830-5757-48a7-8060-af7d189243df", "address": "fa:16:3e:18:51:2f", "network": {"id": "7fd7f791-da5d-41c4-b182-2562fd2fbf47", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-684227548-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4cbef4831f2c4e27bf644608817217b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadc8e830-57", "ovs_interfaceid": "adc8e830-5757-48a7-8060-af7d189243df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.166970] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464724, 'name': Rename_Task, 'duration_secs': 0.134575} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.167984] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1027.168274] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-78041674-9d1b-4f2f-a3df-5e4bd21d7858 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.177939] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Waiting for the task: (returnval){ [ 1027.177939] env[61728]: value = "task-464725" [ 1027.177939] env[61728]: _type = "Task" [ 1027.177939] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.201808] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464725, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.219149] env[61728]: DEBUG oslo_concurrency.lockutils [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.219382] env[61728]: DEBUG oslo_concurrency.lockutils [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.219632] env[61728]: DEBUG nova.network.neutron [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1027.268037] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Releasing lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.268037] env[61728]: DEBUG nova.objects.instance [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lazy-loading 'flavor' on Instance uuid 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.341213] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4e72a80a-3202-4518-b3de-2f7c2ada3ebf tempest-ServerDiskConfigTestJSON-674284000 tempest-ServerDiskConfigTestJSON-674284000-project-member] Lock "378ef752-6d47-4861-baf1-064b2a180dcc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.729s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.354587] env[61728]: DEBUG oslo_concurrency.lockutils [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.354587] env[61728]: DEBUG oslo_concurrency.lockutils [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.354587] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a01ca071-e4c6-4576-af40-bd0800f38f65 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.360023] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Releasing lock "refresh_cache-5f248cf6-31a9-49ef-a8a8-218987664af7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.360023] env[61728]: DEBUG nova.compute.manager [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Instance network_info: |[{"id": "adc8e830-5757-48a7-8060-af7d189243df", "address": "fa:16:3e:18:51:2f", "network": {"id": "7fd7f791-da5d-41c4-b182-2562fd2fbf47", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-684227548-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4cbef4831f2c4e27bf644608817217b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadc8e830-57", "ovs_interfaceid": "adc8e830-5757-48a7-8060-af7d189243df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1027.360023] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:51:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'adc8e830-5757-48a7-8060-af7d189243df', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1027.368214] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Creating folder: Project (4cbef4831f2c4e27bf644608817217b6). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1027.383547] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5f378c2e-345d-422c-b057-4c8a67def24f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.386027] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c93c17-ed0d-499c-ab45-9738ace684fa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.391739] env[61728]: DEBUG nova.compute.manager [req-a7618e6e-e92e-4136-90b3-62a38adec54a req-dc0dca0f-ad73-4876-814a-9271e7ada0a5 service nova] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Received event network-changed-adc8e830-5757-48a7-8060-af7d189243df {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1027.392023] env[61728]: DEBUG nova.compute.manager [req-a7618e6e-e92e-4136-90b3-62a38adec54a req-dc0dca0f-ad73-4876-814a-9271e7ada0a5 service nova] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Refreshing instance network info cache due to event network-changed-adc8e830-5757-48a7-8060-af7d189243df. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1027.392315] env[61728]: DEBUG oslo_concurrency.lockutils [req-a7618e6e-e92e-4136-90b3-62a38adec54a req-dc0dca0f-ad73-4876-814a-9271e7ada0a5 service nova] Acquiring lock "refresh_cache-5f248cf6-31a9-49ef-a8a8-218987664af7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.392727] env[61728]: DEBUG oslo_concurrency.lockutils [req-a7618e6e-e92e-4136-90b3-62a38adec54a req-dc0dca0f-ad73-4876-814a-9271e7ada0a5 service nova] Acquired lock "refresh_cache-5f248cf6-31a9-49ef-a8a8-218987664af7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.392963] env[61728]: DEBUG nova.network.neutron [req-a7618e6e-e92e-4136-90b3-62a38adec54a req-dc0dca0f-ad73-4876-814a-9271e7ada0a5 service nova] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Refreshing network info cache for port adc8e830-5757-48a7-8060-af7d189243df {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1027.420450] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Reconfiguring VM to detach interface {{(pid=61728) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1027.423042] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1474ef8c-ac96-470b-879f-2394ccbb287c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.437421] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Created folder: Project (4cbef4831f2c4e27bf644608817217b6) in parent group-v121913. [ 1027.437745] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Creating folder: Instances. Parent ref: group-v122187. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1027.438621] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8eb8a356-07df-45be-9a96-15b2163b87b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.446437] env[61728]: DEBUG oslo_vmware.api [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 1027.446437] env[61728]: value = "task-464728" [ 1027.446437] env[61728]: _type = "Task" [ 1027.446437] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.450977] env[61728]: DEBUG nova.network.neutron [-] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.454863] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Created folder: Instances in parent group-v122187. [ 1027.454863] env[61728]: DEBUG oslo.service.loopingcall [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.454863] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1027.454863] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89a46d48-94f3-4099-9e93-014637dbe595 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.475192] env[61728]: INFO nova.compute.manager [-] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Took 1.34 seconds to deallocate network for instance. [ 1027.476062] env[61728]: DEBUG oslo_vmware.api [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464728, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.483841] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1027.483841] env[61728]: value = "task-464729" [ 1027.483841] env[61728]: _type = "Task" [ 1027.483841] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.493972] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464729, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.689579] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "bb7b38d5-53c7-4c05-96f8-9121171d3c46" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.690922] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "bb7b38d5-53c7-4c05-96f8-9121171d3c46" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.690922] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "bb7b38d5-53c7-4c05-96f8-9121171d3c46-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.690922] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "bb7b38d5-53c7-4c05-96f8-9121171d3c46-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.690922] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "bb7b38d5-53c7-4c05-96f8-9121171d3c46-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.693640] env[61728]: DEBUG oslo_vmware.api [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464725, 'name': PowerOnVM_Task, 'duration_secs': 0.45669} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.693640] env[61728]: INFO nova.compute.manager [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Terminating instance [ 1027.694714] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1027.694714] env[61728]: INFO nova.compute.manager [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Took 6.89 seconds to spawn the instance on the hypervisor. [ 1027.694714] env[61728]: DEBUG nova.compute.manager [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1027.695093] env[61728]: DEBUG nova.compute.manager [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1027.695306] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1027.696220] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c982a2f-7bc5-49ae-ab8f-9eabbfde86cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.698911] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c73626c-6005-4135-85dc-f71fa65723b6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.709520] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1027.709954] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96377ac6-a3be-4508-bf8d-f215b9d0f1dd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.716114] env[61728]: DEBUG oslo_vmware.api [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1027.716114] env[61728]: value = "task-464730" [ 1027.716114] env[61728]: _type = "Task" [ 1027.716114] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.726262] env[61728]: DEBUG oslo_vmware.api [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464730, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.773529] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc00e8e-0ef1-4ce9-8191-34eb35b33bbe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.800205] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1027.803065] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63dce326-e493-488d-ba17-9fa0db706512 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.814515] env[61728]: DEBUG oslo_vmware.api [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1027.814515] env[61728]: value = "task-464731" [ 1027.814515] env[61728]: _type = "Task" [ 1027.814515] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.824605] env[61728]: DEBUG oslo_vmware.api [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464731, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.962085] env[61728]: DEBUG oslo_vmware.api [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464728, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.995828] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464729, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.027009] env[61728]: INFO nova.compute.manager [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Took 0.55 seconds to detach 1 volumes for instance. [ 1028.041155] env[61728]: DEBUG nova.network.neutron [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance_info_cache with network_info: [{"id": "4660e4da-30e8-4e51-9aea-a9375732537f", "address": "fa:16:3e:e1:e9:0c", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4660e4da-30", "ovs_interfaceid": "4660e4da-30e8-4e51-9aea-a9375732537f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.088206] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f5878eb-4bd5-4ce3-9fad-d69532d01007 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.101406] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6d1102-5c86-400c-85ed-908c6d8e3c0f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.139935] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293f5942-0ff0-4586-9fe4-9cb45fdf90a8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.153140] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72afe0a-861f-4eba-bcf5-9edf1445c6bd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.169325] env[61728]: DEBUG nova.compute.provider_tree [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.189626] env[61728]: DEBUG nova.network.neutron [req-a7618e6e-e92e-4136-90b3-62a38adec54a req-dc0dca0f-ad73-4876-814a-9271e7ada0a5 service nova] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Updated VIF entry in instance network info cache for port adc8e830-5757-48a7-8060-af7d189243df. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1028.190094] env[61728]: DEBUG nova.network.neutron [req-a7618e6e-e92e-4136-90b3-62a38adec54a req-dc0dca0f-ad73-4876-814a-9271e7ada0a5 service nova] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Updating instance_info_cache with network_info: [{"id": "adc8e830-5757-48a7-8060-af7d189243df", "address": "fa:16:3e:18:51:2f", "network": {"id": "7fd7f791-da5d-41c4-b182-2562fd2fbf47", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-684227548-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4cbef4831f2c4e27bf644608817217b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadc8e830-57", "ovs_interfaceid": "adc8e830-5757-48a7-8060-af7d189243df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.220495] env[61728]: INFO nova.compute.manager [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Took 21.37 seconds to build instance. [ 1028.228257] env[61728]: DEBUG oslo_vmware.api [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464730, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.325604] env[61728]: DEBUG oslo_vmware.api [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464731, 'name': PowerOffVM_Task, 'duration_secs': 0.240637} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.325900] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1028.331874] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Reconfiguring VM instance instance-0000005c to detach disk 2002 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1028.332278] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae88d1b9-8402-4165-9184-bd1cfe432d74 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.350792] env[61728]: DEBUG oslo_vmware.api [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1028.350792] env[61728]: value = "task-464732" [ 1028.350792] env[61728]: _type = "Task" [ 1028.350792] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.359488] env[61728]: DEBUG oslo_vmware.api [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464732, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.462066] env[61728]: DEBUG oslo_vmware.api [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464728, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.496639] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464729, 'name': CreateVM_Task, 'duration_secs': 0.746887} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.496821] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1028.497522] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.497693] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.498033] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1028.498518] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e1ff2d1-230d-43e5-a712-0ac2efaf1c27 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.503010] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Waiting for the task: (returnval){ [ 1028.503010] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c00e4d-8f15-03b3-dcae-0f65324021ee" [ 1028.503010] env[61728]: _type = "Task" [ 1028.503010] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.510251] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c00e4d-8f15-03b3-dcae-0f65324021ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.541535] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.547225] env[61728]: DEBUG oslo_concurrency.lockutils [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.551887] env[61728]: DEBUG nova.compute.manager [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1028.552317] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31943959-ec4f-403a-b755-edb010922de0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.674629] env[61728]: DEBUG nova.scheduler.client.report [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.693520] env[61728]: DEBUG oslo_concurrency.lockutils [req-a7618e6e-e92e-4136-90b3-62a38adec54a req-dc0dca0f-ad73-4876-814a-9271e7ada0a5 service nova] Releasing lock "refresh_cache-5f248cf6-31a9-49ef-a8a8-218987664af7" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.694277] env[61728]: DEBUG nova.compute.manager [req-a7618e6e-e92e-4136-90b3-62a38adec54a req-dc0dca0f-ad73-4876-814a-9271e7ada0a5 service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Received event network-vif-deleted-83657468-5cac-4258-84f4-b105395d9cfa {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1028.694277] env[61728]: INFO nova.compute.manager [req-a7618e6e-e92e-4136-90b3-62a38adec54a req-dc0dca0f-ad73-4876-814a-9271e7ada0a5 service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Neutron deleted interface 83657468-5cac-4258-84f4-b105395d9cfa; detaching it from the instance and deleting it from the info cache [ 1028.694277] env[61728]: DEBUG nova.network.neutron [req-a7618e6e-e92e-4136-90b3-62a38adec54a req-dc0dca0f-ad73-4876-814a-9271e7ada0a5 service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.722990] env[61728]: DEBUG oslo_concurrency.lockutils [None req-39a28007-451c-4747-8fbc-bcc2d589eca0 tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Lock "9efdee89-4c88-4af3-af04-5281e8f6bc25" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.882s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.726597] env[61728]: DEBUG oslo_vmware.api [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464730, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.744073] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.744310] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.860961] env[61728]: DEBUG oslo_vmware.api [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464732, 'name': ReconfigVM_Task, 'duration_secs': 0.232273} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.861257] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Reconfigured VM instance instance-0000005c to detach disk 2002 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1028.861540] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1028.861834] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ffbc2d6-359e-4dfe-9395-3807b5517c79 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.868547] env[61728]: DEBUG oslo_vmware.api [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1028.868547] env[61728]: value = "task-464733" [ 1028.868547] env[61728]: _type = "Task" [ 1028.868547] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.876407] env[61728]: DEBUG oslo_vmware.api [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464733, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.962191] env[61728]: DEBUG oslo_vmware.api [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464728, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.971055] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Acquiring lock "9efdee89-4c88-4af3-af04-5281e8f6bc25" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.971877] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Lock "9efdee89-4c88-4af3-af04-5281e8f6bc25" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.971877] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Acquiring lock "9efdee89-4c88-4af3-af04-5281e8f6bc25-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.972063] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Lock "9efdee89-4c88-4af3-af04-5281e8f6bc25-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.972714] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Lock "9efdee89-4c88-4af3-af04-5281e8f6bc25-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.974963] env[61728]: INFO nova.compute.manager [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Terminating instance [ 1028.976936] env[61728]: DEBUG nova.compute.manager [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1028.977165] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1028.978037] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1f82d9-79bd-41c1-9e1b-34e70522e4fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.986255] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1028.986534] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fd6e318-e7ff-4fad-83ad-967f5622224b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.992745] env[61728]: DEBUG oslo_vmware.api [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Waiting for the task: (returnval){ [ 1028.992745] env[61728]: value = "task-464734" [ 1028.992745] env[61728]: _type = "Task" [ 1028.992745] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.001209] env[61728]: DEBUG oslo_vmware.api [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464734, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.013525] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c00e4d-8f15-03b3-dcae-0f65324021ee, 'name': SearchDatastore_Task, 'duration_secs': 0.009124} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.013866] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.014132] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.014374] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.014548] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.014742] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1029.015086] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-126dcb0d-4425-4a79-b6f8-5253226154a4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.022623] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1029.022810] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1029.023547] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ffdb541-dfc4-4722-8167-e273d9a6ac83 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.028392] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Waiting for the task: (returnval){ [ 1029.028392] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52186b64-8ae2-1f5e-0939-508bf4fa6e56" [ 1029.028392] env[61728]: _type = "Task" [ 1029.028392] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.036074] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52186b64-8ae2-1f5e-0939-508bf4fa6e56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.180189] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.180774] env[61728]: DEBUG nova.compute.manager [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1029.187879] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.367s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.189299] env[61728]: INFO nova.compute.claims [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1029.196111] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-487d94d6-ea15-4ad4-a1d5-7157ab70a652 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.205786] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd1ee2b-4075-47e6-a80e-8dca889fa17c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.229205] env[61728]: DEBUG oslo_vmware.api [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464730, 'name': PowerOffVM_Task, 'duration_secs': 1.250383} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.230500] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1029.230500] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1029.230770] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae5ad256-f27d-4fca-b24f-9ac181fa7a40 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.246529] env[61728]: DEBUG nova.compute.manager [req-a7618e6e-e92e-4136-90b3-62a38adec54a req-dc0dca0f-ad73-4876-814a-9271e7ada0a5 service nova] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Detach interface failed, port_id=83657468-5cac-4258-84f4-b105395d9cfa, reason: Instance f9f07464-9f43-43fd-8895-2673861747fa could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1029.247141] env[61728]: DEBUG nova.compute.manager [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1029.288787] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1029.289421] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1029.289421] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleting the datastore file [datastore1] bb7b38d5-53c7-4c05-96f8-9121171d3c46 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1029.289421] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8ae48ade-6970-4b90-883f-77c6df9911ee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.296634] env[61728]: DEBUG oslo_vmware.api [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1029.296634] env[61728]: value = "task-464736" [ 1029.296634] env[61728]: _type = "Task" [ 1029.296634] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.304650] env[61728]: DEBUG oslo_vmware.api [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464736, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.379102] env[61728]: DEBUG oslo_vmware.api [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464733, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.461763] env[61728]: DEBUG oslo_vmware.api [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464728, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.503219] env[61728]: DEBUG oslo_vmware.api [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464734, 'name': PowerOffVM_Task, 'duration_secs': 0.187889} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.503219] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1029.503219] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1029.503763] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3eccf8db-a64e-428c-88cf-6ef31f4e74fe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.538570] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52186b64-8ae2-1f5e-0939-508bf4fa6e56, 'name': SearchDatastore_Task, 'duration_secs': 0.012007} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.539402] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83be7528-1c12-42a6-949a-688394a7b745 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.545282] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Waiting for the task: (returnval){ [ 1029.545282] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5229a078-ba80-4dd0-680a-c1ff0c19c615" [ 1029.545282] env[61728]: _type = "Task" [ 1029.545282] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.554233] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5229a078-ba80-4dd0-680a-c1ff0c19c615, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.568793] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8723156b-cbfa-497d-b63c-fe9011f19722 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.571772] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1029.572059] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1029.572192] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Deleting the datastore file [datastore1] 9efdee89-4c88-4af3-af04-5281e8f6bc25 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1029.572436] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dcce98ea-022f-4c85-a4c5-9476d2df71f0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.582136] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Doing hard reboot of VM {{(pid=61728) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 1029.583450] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-c1c05abb-70e0-422e-8d32-348863986e9c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.584959] env[61728]: DEBUG oslo_vmware.api [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Waiting for the task: (returnval){ [ 1029.584959] env[61728]: value = "task-464738" [ 1029.584959] env[61728]: _type = "Task" [ 1029.584959] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.589456] env[61728]: DEBUG oslo_vmware.api [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1029.589456] env[61728]: value = "task-464739" [ 1029.589456] env[61728]: _type = "Task" [ 1029.589456] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.596530] env[61728]: DEBUG oslo_vmware.api [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464738, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.601160] env[61728]: DEBUG oslo_vmware.api [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464739, 'name': ResetVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.694099] env[61728]: DEBUG nova.compute.utils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1029.699777] env[61728]: DEBUG nova.compute.manager [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1029.699983] env[61728]: DEBUG nova.network.neutron [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1029.777763] env[61728]: DEBUG nova.policy [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e17b05d01da49e68afe0b7da5db6fb9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '833a646ea21a48c4951ed8cedebaa98f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1029.780716] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.806613] env[61728]: DEBUG oslo_vmware.api [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464736, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.336252} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.806884] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.807094] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1029.807283] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1029.807467] env[61728]: INFO nova.compute.manager [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Took 2.11 seconds to destroy the instance on the hypervisor. [ 1029.807717] env[61728]: DEBUG oslo.service.loopingcall [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.807947] env[61728]: DEBUG nova.compute.manager [-] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1029.808075] env[61728]: DEBUG nova.network.neutron [-] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1029.879747] env[61728]: DEBUG oslo_vmware.api [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464733, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.966514] env[61728]: DEBUG oslo_vmware.api [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464728, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.056102] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5229a078-ba80-4dd0-680a-c1ff0c19c615, 'name': SearchDatastore_Task, 'duration_secs': 0.019313} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.056634] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.057048] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 5f248cf6-31a9-49ef-a8a8-218987664af7/5f248cf6-31a9-49ef-a8a8-218987664af7.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1030.057722] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5b60fd1-2014-4091-8892-afdcb5b7aec7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.063996] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Waiting for the task: (returnval){ [ 1030.063996] env[61728]: value = "task-464740" [ 1030.063996] env[61728]: _type = "Task" [ 1030.063996] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.071978] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464740, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.093827] env[61728]: DEBUG oslo_vmware.api [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Task: {'id': task-464738, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.389903} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.097484] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1030.097870] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1030.098180] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1030.098503] env[61728]: INFO nova.compute.manager [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1030.098850] env[61728]: DEBUG oslo.service.loopingcall [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.099178] env[61728]: DEBUG nova.compute.manager [-] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1030.099369] env[61728]: DEBUG nova.network.neutron [-] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1030.106149] env[61728]: DEBUG oslo_vmware.api [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464739, 'name': ResetVM_Task, 'duration_secs': 0.098347} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.106923] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Did hard reboot of VM {{(pid=61728) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 1030.107265] env[61728]: DEBUG nova.compute.manager [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1030.108116] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc83bf23-35b3-4db3-96e5-558ff7390477 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.204703] env[61728]: DEBUG nova.compute.manager [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1030.386610] env[61728]: DEBUG oslo_vmware.api [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464733, 'name': PowerOnVM_Task, 'duration_secs': 1.150358} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.390254] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1030.390556] env[61728]: DEBUG nova.compute.manager [None req-0414fcbd-b57d-4ccc-a1bc-7f6cd6cd2dd1 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1030.394759] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183001d7-232f-499d-9ca9-2e545030b420 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.442594] env[61728]: DEBUG nova.network.neutron [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Successfully created port: 4155b1a7-975f-404c-8374-47665c8046ca {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1030.468992] env[61728]: DEBUG oslo_vmware.api [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464728, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.520023] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4630b7d8-e307-489f-82fd-adf285889f31 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.530146] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b77051c-4303-44c0-bd40-b574eb6f2c84 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.573761] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109fa033-46c8-4e34-b84b-c83ee92e5933 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.579043] env[61728]: DEBUG nova.compute.manager [req-5a80d195-1451-4d32-9572-6c409f831591 req-a9fcaa22-4553-4f99-a45c-3414826109b4 service nova] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Received event network-vif-deleted-216193d1-1332-49ba-8702-dacc02486ec1 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1030.579043] env[61728]: INFO nova.compute.manager [req-5a80d195-1451-4d32-9572-6c409f831591 req-a9fcaa22-4553-4f99-a45c-3414826109b4 service nova] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Neutron deleted interface 216193d1-1332-49ba-8702-dacc02486ec1; detaching it from the instance and deleting it from the info cache [ 1030.579043] env[61728]: DEBUG nova.network.neutron [req-5a80d195-1451-4d32-9572-6c409f831591 req-a9fcaa22-4553-4f99-a45c-3414826109b4 service nova] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.590234] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e570ee8-4388-4348-9673-c534421dd718 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.596034] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464740, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.609618] env[61728]: DEBUG nova.compute.provider_tree [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.612689] env[61728]: DEBUG nova.compute.manager [req-b9e0604b-f435-4ac0-a49d-d5203ca05dd1 req-fac1493b-7837-4277-92a0-2e46b72d4518 service nova] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Received event network-vif-deleted-1db8e5be-43b2-4fa8-954e-dce8e21ff0d4 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1030.612689] env[61728]: INFO nova.compute.manager [req-b9e0604b-f435-4ac0-a49d-d5203ca05dd1 req-fac1493b-7837-4277-92a0-2e46b72d4518 service nova] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Neutron deleted interface 1db8e5be-43b2-4fa8-954e-dce8e21ff0d4; detaching it from the instance and deleting it from the info cache [ 1030.612689] env[61728]: DEBUG nova.network.neutron [req-b9e0604b-f435-4ac0-a49d-d5203ca05dd1 req-fac1493b-7837-4277-92a0-2e46b72d4518 service nova] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.624056] env[61728]: DEBUG oslo_concurrency.lockutils [None req-53a51d38-41fd-4bce-b802-d6a8b504440e tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.942s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.872610] env[61728]: DEBUG nova.network.neutron [-] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.887051] env[61728]: DEBUG nova.network.neutron [-] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.963727] env[61728]: DEBUG oslo_vmware.api [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464728, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.079082] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464740, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.732658} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.079359] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 5f248cf6-31a9-49ef-a8a8-218987664af7/5f248cf6-31a9-49ef-a8a8-218987664af7.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1031.079617] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1031.079872] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0c13b5c3-60cc-4fe6-8d6d-1f74b9015faf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.081966] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ac6b7cc-91b3-4689-8e33-0120f09b66b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.100153] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb279e6-633e-4ab5-9f14-7cdefc4abff2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.111540] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Waiting for the task: (returnval){ [ 1031.111540] env[61728]: value = "task-464741" [ 1031.111540] env[61728]: _type = "Task" [ 1031.111540] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.116133] env[61728]: DEBUG nova.scheduler.client.report [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1031.119309] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f666f649-3e35-4925-84b4-53a7c39780ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.133080] env[61728]: DEBUG nova.compute.manager [req-5a80d195-1451-4d32-9572-6c409f831591 req-a9fcaa22-4553-4f99-a45c-3414826109b4 service nova] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Detach interface failed, port_id=216193d1-1332-49ba-8702-dacc02486ec1, reason: Instance 9efdee89-4c88-4af3-af04-5281e8f6bc25 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1031.138719] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464741, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.141672] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f4d838e-fb46-4361-b16a-a602aa5baaef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.174349] env[61728]: DEBUG nova.compute.manager [req-b9e0604b-f435-4ac0-a49d-d5203ca05dd1 req-fac1493b-7837-4277-92a0-2e46b72d4518 service nova] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Detach interface failed, port_id=1db8e5be-43b2-4fa8-954e-dce8e21ff0d4, reason: Instance bb7b38d5-53c7-4c05-96f8-9121171d3c46 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1031.216957] env[61728]: DEBUG nova.compute.manager [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1031.243488] env[61728]: DEBUG nova.virt.hardware [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.243868] env[61728]: DEBUG nova.virt.hardware [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.243952] env[61728]: DEBUG nova.virt.hardware [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.244165] env[61728]: DEBUG nova.virt.hardware [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.244334] env[61728]: DEBUG nova.virt.hardware [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.244499] env[61728]: DEBUG nova.virt.hardware [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.244720] env[61728]: DEBUG nova.virt.hardware [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.245069] env[61728]: DEBUG nova.virt.hardware [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.245069] env[61728]: DEBUG nova.virt.hardware [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.245245] env[61728]: DEBUG nova.virt.hardware [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.245500] env[61728]: DEBUG nova.virt.hardware [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.246311] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6e6a9e-5e88-449e-9b41-5bca5d8be4d2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.254828] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37cb8bd-cb58-40b2-be85-764c15193fd7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.375508] env[61728]: INFO nova.compute.manager [-] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Took 1.28 seconds to deallocate network for instance. [ 1031.389621] env[61728]: INFO nova.compute.manager [-] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Took 1.58 seconds to deallocate network for instance. [ 1031.465307] env[61728]: DEBUG oslo_vmware.api [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464728, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.623008] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.623602] env[61728]: DEBUG nova.compute.manager [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1031.629144] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464741, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111247} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.629144] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.086s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.629144] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.629488] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.849s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.632017] env[61728]: INFO nova.compute.claims [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1031.634350] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1031.636012] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86f5da5-e3fb-4468-af64-f7e47f27d3b4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.669368] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 5f248cf6-31a9-49ef-a8a8-218987664af7/5f248cf6-31a9-49ef-a8a8-218987664af7.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.670455] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3d00e50-f1cc-4389-89df-b384057e9b6e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.688623] env[61728]: INFO nova.scheduler.client.report [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleted allocations for instance f9f07464-9f43-43fd-8895-2673861747fa [ 1031.698628] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Waiting for the task: (returnval){ [ 1031.698628] env[61728]: value = "task-464742" [ 1031.698628] env[61728]: _type = "Task" [ 1031.698628] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.709529] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464742, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.883159] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.896395] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.965575] env[61728]: DEBUG oslo_vmware.api [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464728, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.129048] env[61728]: DEBUG nova.compute.utils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1032.130705] env[61728]: DEBUG nova.compute.manager [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1032.130888] env[61728]: DEBUG nova.network.neutron [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1032.207217] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0b3f0a39-5aeb-42b5-a6a4-268066f60e9b tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "f9f07464-9f43-43fd-8895-2673861747fa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.922s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.221601] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464742, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.242176] env[61728]: DEBUG nova.policy [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd32b5118a94042f797dea1fdf6bd05b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08367aaceba548fe93faaedf6371817d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1032.258851] env[61728]: DEBUG nova.network.neutron [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Successfully updated port: 4155b1a7-975f-404c-8374-47665c8046ca {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1032.443534] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.443777] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.468740] env[61728]: DEBUG oslo_vmware.api [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464728, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.596457] env[61728]: DEBUG nova.compute.manager [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Received event network-changed-67077c45-d87f-4d2d-a48b-d23a96d35f8a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1032.596805] env[61728]: DEBUG nova.compute.manager [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Refreshing instance network info cache due to event network-changed-67077c45-d87f-4d2d-a48b-d23a96d35f8a. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1032.597153] env[61728]: DEBUG oslo_concurrency.lockutils [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] Acquiring lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.597414] env[61728]: DEBUG oslo_concurrency.lockutils [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] Acquired lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.599632] env[61728]: DEBUG nova.network.neutron [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Refreshing network info cache for port 67077c45-d87f-4d2d-a48b-d23a96d35f8a {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1032.636220] env[61728]: DEBUG nova.compute.manager [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1032.646027] env[61728]: DEBUG nova.compute.manager [req-95f11cb9-06c5-4d4f-8ef6-f9887f32d951 req-9658bb57-0f54-43ab-a763-0f18d3a53968 service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Received event network-vif-plugged-4155b1a7-975f-404c-8374-47665c8046ca {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1032.646027] env[61728]: DEBUG oslo_concurrency.lockutils [req-95f11cb9-06c5-4d4f-8ef6-f9887f32d951 req-9658bb57-0f54-43ab-a763-0f18d3a53968 service nova] Acquiring lock "d55beb02-00fc-4df6-a239-9e5d776a7c18-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.646027] env[61728]: DEBUG oslo_concurrency.lockutils [req-95f11cb9-06c5-4d4f-8ef6-f9887f32d951 req-9658bb57-0f54-43ab-a763-0f18d3a53968 service nova] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.646027] env[61728]: DEBUG oslo_concurrency.lockutils [req-95f11cb9-06c5-4d4f-8ef6-f9887f32d951 req-9658bb57-0f54-43ab-a763-0f18d3a53968 service nova] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.646027] env[61728]: DEBUG nova.compute.manager [req-95f11cb9-06c5-4d4f-8ef6-f9887f32d951 req-9658bb57-0f54-43ab-a763-0f18d3a53968 service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] No waiting events found dispatching network-vif-plugged-4155b1a7-975f-404c-8374-47665c8046ca {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1032.646027] env[61728]: WARNING nova.compute.manager [req-95f11cb9-06c5-4d4f-8ef6-f9887f32d951 req-9658bb57-0f54-43ab-a763-0f18d3a53968 service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Received unexpected event network-vif-plugged-4155b1a7-975f-404c-8374-47665c8046ca for instance with vm_state building and task_state spawning. [ 1032.646027] env[61728]: DEBUG nova.compute.manager [req-95f11cb9-06c5-4d4f-8ef6-f9887f32d951 req-9658bb57-0f54-43ab-a763-0f18d3a53968 service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Received event network-changed-4155b1a7-975f-404c-8374-47665c8046ca {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1032.646027] env[61728]: DEBUG nova.compute.manager [req-95f11cb9-06c5-4d4f-8ef6-f9887f32d951 req-9658bb57-0f54-43ab-a763-0f18d3a53968 service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Refreshing instance network info cache due to event network-changed-4155b1a7-975f-404c-8374-47665c8046ca. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1032.646027] env[61728]: DEBUG oslo_concurrency.lockutils [req-95f11cb9-06c5-4d4f-8ef6-f9887f32d951 req-9658bb57-0f54-43ab-a763-0f18d3a53968 service nova] Acquiring lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.646027] env[61728]: DEBUG oslo_concurrency.lockutils [req-95f11cb9-06c5-4d4f-8ef6-f9887f32d951 req-9658bb57-0f54-43ab-a763-0f18d3a53968 service nova] Acquired lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.646027] env[61728]: DEBUG nova.network.neutron [req-95f11cb9-06c5-4d4f-8ef6-f9887f32d951 req-9658bb57-0f54-43ab-a763-0f18d3a53968 service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Refreshing network info cache for port 4155b1a7-975f-404c-8374-47665c8046ca {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1032.709577] env[61728]: DEBUG nova.network.neutron [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Successfully created port: e7f1c36b-89f0-4858-96a0-3c1875c4bb29 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1032.723160] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464742, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.763906] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.913950] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23838e5-0155-4385-b54c-d32635daa87d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.923931] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a07ae3-7ef7-45b9-a3a2-0de83b023ec3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.975804] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42ff2d3-1d2e-4d96-8736-2a579b11cab9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.979906] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.979984] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Starting heal instance info cache {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1032.980092] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Rebuilding the list of instances to heal {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1032.990348] env[61728]: DEBUG oslo_vmware.api [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464728, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.992552] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72fb8ef-b804-44a7-9c86-178036c116f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.011067] env[61728]: DEBUG nova.compute.provider_tree [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.193378] env[61728]: DEBUG nova.network.neutron [req-95f11cb9-06c5-4d4f-8ef6-f9887f32d951 req-9658bb57-0f54-43ab-a763-0f18d3a53968 service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1033.213836] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464742, 'name': ReconfigVM_Task, 'duration_secs': 1.050649} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.219274] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 5f248cf6-31a9-49ef-a8a8-218987664af7/5f248cf6-31a9-49ef-a8a8-218987664af7.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.220456] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ddf5ea33-096e-46dc-bbb1-c219fff31bdd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.230124] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Waiting for the task: (returnval){ [ 1033.230124] env[61728]: value = "task-464743" [ 1033.230124] env[61728]: _type = "Task" [ 1033.230124] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.245053] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464743, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.285289] env[61728]: DEBUG nova.network.neutron [req-95f11cb9-06c5-4d4f-8ef6-f9887f32d951 req-9658bb57-0f54-43ab-a763-0f18d3a53968 service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.374199] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.374461] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.479930] env[61728]: DEBUG oslo_vmware.api [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464728, 'name': ReconfigVM_Task, 'duration_secs': 5.770613} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.480270] env[61728]: DEBUG oslo_concurrency.lockutils [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.480501] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Reconfigured VM to detach interface {{(pid=61728) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1033.488027] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Skipping network cache update for instance because it is Building. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1033.488027] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Skipping network cache update for instance because it is Building. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1033.488027] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Skipping network cache update for instance because it is Building. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1033.488027] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Skipping network cache update for instance because it is Building. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1033.515038] env[61728]: DEBUG nova.scheduler.client.report [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1033.540721] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "refresh_cache-40ef29df-12d6-4274-846a-4a2c32752b8b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.540874] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquired lock "refresh_cache-40ef29df-12d6-4274-846a-4a2c32752b8b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.541388] env[61728]: DEBUG nova.network.neutron [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Forcefully refreshing network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1033.541388] env[61728]: DEBUG nova.objects.instance [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lazy-loading 'info_cache' on Instance uuid 40ef29df-12d6-4274-846a-4a2c32752b8b {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.651821] env[61728]: DEBUG nova.network.neutron [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Updated VIF entry in instance network info cache for port 67077c45-d87f-4d2d-a48b-d23a96d35f8a. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1033.651821] env[61728]: DEBUG nova.network.neutron [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Updating instance_info_cache with network_info: [{"id": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "address": "fa:16:3e:d1:31:88", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67077c45-d8", "ovs_interfaceid": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.660112] env[61728]: DEBUG nova.compute.manager [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1033.695556] env[61728]: DEBUG nova.virt.hardware [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1033.695820] env[61728]: DEBUG nova.virt.hardware [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1033.695983] env[61728]: DEBUG nova.virt.hardware [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1033.696224] env[61728]: DEBUG nova.virt.hardware [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1033.696427] env[61728]: DEBUG nova.virt.hardware [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1033.696595] env[61728]: DEBUG nova.virt.hardware [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1033.696881] env[61728]: DEBUG nova.virt.hardware [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1033.699154] env[61728]: DEBUG nova.virt.hardware [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1033.699154] env[61728]: DEBUG nova.virt.hardware [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1033.699281] env[61728]: DEBUG nova.virt.hardware [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1033.699422] env[61728]: DEBUG nova.virt.hardware [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1033.700321] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df1a80db-835e-4e94-8710-9dbf6c12db62 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.708691] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb7fdce-c959-46dd-bacd-db8747a4f4ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.738965] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464743, 'name': Rename_Task, 'duration_secs': 0.183365} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.739276] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1033.739525] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6c3fcded-47fc-4730-8056-89f4df31b9a3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.745557] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Waiting for the task: (returnval){ [ 1033.745557] env[61728]: value = "task-464744" [ 1033.745557] env[61728]: _type = "Task" [ 1033.745557] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.753186] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464744, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.789286] env[61728]: DEBUG oslo_concurrency.lockutils [req-95f11cb9-06c5-4d4f-8ef6-f9887f32d951 req-9658bb57-0f54-43ab-a763-0f18d3a53968 service nova] Releasing lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.789706] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquired lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.789872] env[61728]: DEBUG nova.network.neutron [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1033.879025] env[61728]: DEBUG nova.compute.manager [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1034.021344] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.022402] env[61728]: DEBUG nova.compute.manager [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1034.027023] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.144s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.027498] env[61728]: DEBUG nova.objects.instance [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Lazy-loading 'resources' on Instance uuid 9efdee89-4c88-4af3-af04-5281e8f6bc25 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.155772] env[61728]: DEBUG oslo_concurrency.lockutils [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] Releasing lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.156172] env[61728]: DEBUG nova.compute.manager [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Received event network-changed-67077c45-d87f-4d2d-a48b-d23a96d35f8a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1034.156437] env[61728]: DEBUG nova.compute.manager [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Refreshing instance network info cache due to event network-changed-67077c45-d87f-4d2d-a48b-d23a96d35f8a. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1034.156798] env[61728]: DEBUG oslo_concurrency.lockutils [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] Acquiring lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.156968] env[61728]: DEBUG oslo_concurrency.lockutils [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] Acquired lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.157239] env[61728]: DEBUG nova.network.neutron [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Refreshing network info cache for port 67077c45-d87f-4d2d-a48b-d23a96d35f8a {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1034.261416] env[61728]: DEBUG oslo_vmware.api [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464744, 'name': PowerOnVM_Task, 'duration_secs': 0.510939} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.263081] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1034.263318] env[61728]: INFO nova.compute.manager [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Took 9.23 seconds to spawn the instance on the hypervisor. [ 1034.263513] env[61728]: DEBUG nova.compute.manager [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.264374] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04759d8-6762-43de-b64d-e9352fc1d13f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.325028] env[61728]: DEBUG nova.network.neutron [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1034.405221] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.432142] env[61728]: DEBUG nova.network.neutron [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Successfully updated port: e7f1c36b-89f0-4858-96a0-3c1875c4bb29 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1034.533023] env[61728]: DEBUG nova.compute.utils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1034.535779] env[61728]: DEBUG nova.compute.manager [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1034.536136] env[61728]: DEBUG nova.network.neutron [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1034.604466] env[61728]: DEBUG nova.policy [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3baf998a00e494ba33dab6a0a2a88e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '836843da5be34c649d9a48a83e658288', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1034.619026] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "705c8838-f605-46fa-b036-ec212f673f66" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.619707] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "705c8838-f605-46fa-b036-ec212f673f66" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.620128] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "705c8838-f605-46fa-b036-ec212f673f66-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.620519] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "705c8838-f605-46fa-b036-ec212f673f66-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.620838] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "705c8838-f605-46fa-b036-ec212f673f66-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.624043] env[61728]: INFO nova.compute.manager [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Terminating instance [ 1034.627047] env[61728]: DEBUG nova.compute.manager [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1034.628092] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1034.629113] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65d7898-68f8-4154-8f36-adfa7bbbadd3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.641776] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1034.642921] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e9baa853-e139-450e-8cbd-2f6eb1ba2c9b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.650724] env[61728]: DEBUG oslo_vmware.api [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 1034.650724] env[61728]: value = "task-464745" [ 1034.650724] env[61728]: _type = "Task" [ 1034.650724] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.663683] env[61728]: DEBUG oslo_vmware.api [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464745, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.779498] env[61728]: DEBUG nova.compute.manager [req-4547e6bc-bab0-4da4-bf41-795e4c3b05d8 req-ee50b4ca-afbe-436a-98e1-03a371d138f2 service nova] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Received event network-vif-plugged-e7f1c36b-89f0-4858-96a0-3c1875c4bb29 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1034.779498] env[61728]: DEBUG oslo_concurrency.lockutils [req-4547e6bc-bab0-4da4-bf41-795e4c3b05d8 req-ee50b4ca-afbe-436a-98e1-03a371d138f2 service nova] Acquiring lock "6c154517-68e8-4e70-86eb-cf53f6f812d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.779719] env[61728]: DEBUG oslo_concurrency.lockutils [req-4547e6bc-bab0-4da4-bf41-795e4c3b05d8 req-ee50b4ca-afbe-436a-98e1-03a371d138f2 service nova] Lock "6c154517-68e8-4e70-86eb-cf53f6f812d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.779894] env[61728]: DEBUG oslo_concurrency.lockutils [req-4547e6bc-bab0-4da4-bf41-795e4c3b05d8 req-ee50b4ca-afbe-436a-98e1-03a371d138f2 service nova] Lock "6c154517-68e8-4e70-86eb-cf53f6f812d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.780080] env[61728]: DEBUG nova.compute.manager [req-4547e6bc-bab0-4da4-bf41-795e4c3b05d8 req-ee50b4ca-afbe-436a-98e1-03a371d138f2 service nova] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] No waiting events found dispatching network-vif-plugged-e7f1c36b-89f0-4858-96a0-3c1875c4bb29 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1034.780413] env[61728]: WARNING nova.compute.manager [req-4547e6bc-bab0-4da4-bf41-795e4c3b05d8 req-ee50b4ca-afbe-436a-98e1-03a371d138f2 service nova] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Received unexpected event network-vif-plugged-e7f1c36b-89f0-4858-96a0-3c1875c4bb29 for instance with vm_state building and task_state spawning. [ 1034.780680] env[61728]: DEBUG nova.compute.manager [req-4547e6bc-bab0-4da4-bf41-795e4c3b05d8 req-ee50b4ca-afbe-436a-98e1-03a371d138f2 service nova] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Received event network-changed-e7f1c36b-89f0-4858-96a0-3c1875c4bb29 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1034.780966] env[61728]: DEBUG nova.compute.manager [req-4547e6bc-bab0-4da4-bf41-795e4c3b05d8 req-ee50b4ca-afbe-436a-98e1-03a371d138f2 service nova] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Refreshing instance network info cache due to event network-changed-e7f1c36b-89f0-4858-96a0-3c1875c4bb29. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1034.781317] env[61728]: DEBUG oslo_concurrency.lockutils [req-4547e6bc-bab0-4da4-bf41-795e4c3b05d8 req-ee50b4ca-afbe-436a-98e1-03a371d138f2 service nova] Acquiring lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.781594] env[61728]: DEBUG oslo_concurrency.lockutils [req-4547e6bc-bab0-4da4-bf41-795e4c3b05d8 req-ee50b4ca-afbe-436a-98e1-03a371d138f2 service nova] Acquired lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.781915] env[61728]: DEBUG nova.network.neutron [req-4547e6bc-bab0-4da4-bf41-795e4c3b05d8 req-ee50b4ca-afbe-436a-98e1-03a371d138f2 service nova] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Refreshing network info cache for port e7f1c36b-89f0-4858-96a0-3c1875c4bb29 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1034.794118] env[61728]: INFO nova.compute.manager [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Took 22.53 seconds to build instance. [ 1034.803755] env[61728]: DEBUG nova.network.neutron [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Updating instance_info_cache with network_info: [{"id": "4155b1a7-975f-404c-8374-47665c8046ca", "address": "fa:16:3e:e4:b3:9b", "network": {"id": "72d11e2a-907e-419e-b8a4-53aa39f2566a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1755247352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833a646ea21a48c4951ed8cedebaa98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3836fb52-19c6-4e10-a0ca-f0bca73dc887", "external-id": "nsx-vlan-transportzone-964", "segmentation_id": 964, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4155b1a7-97", "ovs_interfaceid": "4155b1a7-975f-404c-8374-47665c8046ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.855570] env[61728]: DEBUG oslo_concurrency.lockutils [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.855776] env[61728]: DEBUG oslo_concurrency.lockutils [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquired lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.855966] env[61728]: DEBUG nova.network.neutron [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1034.862841] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2861fb4-5cb9-4cf2-8ce3-e0045f4f9ba7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.872434] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquiring lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.872675] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.877593] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba473e0c-855a-473a-abc3-517b21748b43 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.920331] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7618507d-413d-475c-8ee7-b11e39ee8f24 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.935032] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.938972] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89928761-8bdd-4492-b99d-367a649c44d0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.955918] env[61728]: DEBUG nova.compute.provider_tree [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.042023] env[61728]: DEBUG nova.compute.manager [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1035.060188] env[61728]: DEBUG nova.network.neutron [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Updated VIF entry in instance network info cache for port 67077c45-d87f-4d2d-a48b-d23a96d35f8a. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1035.060556] env[61728]: DEBUG nova.network.neutron [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Updating instance_info_cache with network_info: [{"id": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "address": "fa:16:3e:d1:31:88", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67077c45-d8", "ovs_interfaceid": "67077c45-d87f-4d2d-a48b-d23a96d35f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.131276] env[61728]: DEBUG nova.network.neutron [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Successfully created port: 63b2bf95-c628-4c62-bdbd-051192e13a2f {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1035.161526] env[61728]: DEBUG oslo_vmware.api [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464745, 'name': PowerOffVM_Task, 'duration_secs': 0.228042} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.161875] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1035.162077] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1035.162678] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd7eb4d9-01cf-481a-9231-fa74b2b6f9fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.225426] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1035.225426] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1035.225503] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Deleting the datastore file [datastore1] 705c8838-f605-46fa-b036-ec212f673f66 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1035.228245] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9302945-643b-4632-9423-49c743269330 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.235029] env[61728]: DEBUG oslo_vmware.api [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 1035.235029] env[61728]: value = "task-464747" [ 1035.235029] env[61728]: _type = "Task" [ 1035.235029] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.242902] env[61728]: DEBUG oslo_vmware.api [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464747, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.296786] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89b35176-0e59-4d79-8d08-5816141e91db tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Lock "5f248cf6-31a9-49ef-a8a8-218987664af7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.038s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.306333] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Releasing lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.306643] env[61728]: DEBUG nova.compute.manager [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Instance network_info: |[{"id": "4155b1a7-975f-404c-8374-47665c8046ca", "address": "fa:16:3e:e4:b3:9b", "network": {"id": "72d11e2a-907e-419e-b8a4-53aa39f2566a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1755247352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833a646ea21a48c4951ed8cedebaa98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3836fb52-19c6-4e10-a0ca-f0bca73dc887", "external-id": "nsx-vlan-transportzone-964", "segmentation_id": 964, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4155b1a7-97", "ovs_interfaceid": "4155b1a7-975f-404c-8374-47665c8046ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1035.307156] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:b3:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3836fb52-19c6-4e10-a0ca-f0bca73dc887', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4155b1a7-975f-404c-8374-47665c8046ca', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1035.315438] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Creating folder: Project (833a646ea21a48c4951ed8cedebaa98f). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1035.315743] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-630b249e-8541-44d0-9d9b-32a4363b3df1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.327166] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Created folder: Project (833a646ea21a48c4951ed8cedebaa98f) in parent group-v121913. [ 1035.327415] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Creating folder: Instances. Parent ref: group-v122190. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1035.327667] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f4ebbe4-453e-4f08-ab27-0ff89c7ad081 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.336181] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Created folder: Instances in parent group-v122190. [ 1035.336352] env[61728]: DEBUG oslo.service.loopingcall [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.336584] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1035.336834] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c9f6bbae-2835-4a3e-ba32-f007bf5de0c4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.352598] env[61728]: DEBUG nova.network.neutron [req-4547e6bc-bab0-4da4-bf41-795e4c3b05d8 req-ee50b4ca-afbe-436a-98e1-03a371d138f2 service nova] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1035.360030] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1035.360030] env[61728]: value = "task-464750" [ 1035.360030] env[61728]: _type = "Task" [ 1035.360030] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.373348] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464750, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.378054] env[61728]: DEBUG nova.compute.manager [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1035.419429] env[61728]: DEBUG nova.network.neutron [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Updating instance_info_cache with network_info: [{"id": "3f69dcc2-f3d8-4b50-b760-85f6849a6c24", "address": "fa:16:3e:77:1c:7f", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f69dcc2-f3", "ovs_interfaceid": "3f69dcc2-f3d8-4b50-b760-85f6849a6c24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.421668] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Acquiring lock "5f248cf6-31a9-49ef-a8a8-218987664af7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.421969] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Lock "5f248cf6-31a9-49ef-a8a8-218987664af7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.422240] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Acquiring lock "5f248cf6-31a9-49ef-a8a8-218987664af7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.422438] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Lock "5f248cf6-31a9-49ef-a8a8-218987664af7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.422674] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Lock "5f248cf6-31a9-49ef-a8a8-218987664af7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.426930] env[61728]: INFO nova.compute.manager [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Terminating instance [ 1035.428949] env[61728]: DEBUG nova.compute.manager [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1035.429173] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1035.430052] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cdadc38-4ba1-47a1-bc7c-6859ca56f914 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.438914] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1035.439088] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad893f48-9302-4fe0-a944-64c376200c24 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.444424] env[61728]: DEBUG oslo_vmware.api [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Waiting for the task: (returnval){ [ 1035.444424] env[61728]: value = "task-464751" [ 1035.444424] env[61728]: _type = "Task" [ 1035.444424] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.453357] env[61728]: DEBUG oslo_vmware.api [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464751, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.459226] env[61728]: DEBUG nova.scheduler.client.report [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1035.483486] env[61728]: DEBUG nova.network.neutron [req-4547e6bc-bab0-4da4-bf41-795e4c3b05d8 req-ee50b4ca-afbe-436a-98e1-03a371d138f2 service nova] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.567634] env[61728]: DEBUG oslo_concurrency.lockutils [req-24686706-58b3-4a6f-b98f-39092f62ebd7 req-8657065a-a4e0-44ce-906f-00f58ba9d21a service nova] Releasing lock "refresh_cache-4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.677734] env[61728]: INFO nova.network.neutron [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Port d0b341ea-4a0b-46ff-89f2-f96d212d04d9 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1035.678285] env[61728]: DEBUG nova.network.neutron [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Updating instance_info_cache with network_info: [{"id": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "address": "fa:16:3e:57:58:2b", "network": {"id": "20ba9b61-3be9-4ad2-a1d0-88c810873bcb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-277928523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "32bb35e1dfce40e48be08bb568d3f2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec0ae08d-62", "ovs_interfaceid": "ec0ae08d-62b0-4506-bc0b-65e9152dd5e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.744650] env[61728]: DEBUG oslo_vmware.api [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464747, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193898} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.744957] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1035.745116] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1035.745375] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1035.745484] env[61728]: INFO nova.compute.manager [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1035.745732] env[61728]: DEBUG oslo.service.loopingcall [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.745926] env[61728]: DEBUG nova.compute.manager [-] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1035.746037] env[61728]: DEBUG nova.network.neutron [-] [instance: 705c8838-f605-46fa-b036-ec212f673f66] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1035.871245] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464750, 'name': CreateVM_Task, 'duration_secs': 0.314695} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.871245] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1035.871373] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.871635] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.871812] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1035.872067] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6262922b-fefc-410b-80d2-c02cebbb3759 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.876544] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1035.876544] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52791fb2-4486-3c68-9c15-3f05d2838e29" [ 1035.876544] env[61728]: _type = "Task" [ 1035.876544] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.886434] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52791fb2-4486-3c68-9c15-3f05d2838e29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.903642] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.922352] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Releasing lock "refresh_cache-40ef29df-12d6-4274-846a-4a2c32752b8b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.922555] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Updated the network info_cache for instance {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1035.922900] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1035.923157] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1035.923332] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1035.923492] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1035.923647] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1035.923831] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1035.923972] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61728) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1035.924207] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1035.954368] env[61728]: DEBUG oslo_vmware.api [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464751, 'name': PowerOffVM_Task, 'duration_secs': 0.211249} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.954648] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1035.954821] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1035.955092] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23e9ff65-2e4a-4974-8a5f-b27709459b85 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.964212] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.937s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.967153] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.070s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.967153] env[61728]: DEBUG nova.objects.instance [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lazy-loading 'resources' on Instance uuid bb7b38d5-53c7-4c05-96f8-9121171d3c46 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.987168] env[61728]: INFO nova.scheduler.client.report [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Deleted allocations for instance 9efdee89-4c88-4af3-af04-5281e8f6bc25 [ 1035.988798] env[61728]: DEBUG oslo_concurrency.lockutils [req-4547e6bc-bab0-4da4-bf41-795e4c3b05d8 req-ee50b4ca-afbe-436a-98e1-03a371d138f2 service nova] Releasing lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.989173] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.989333] env[61728]: DEBUG nova.network.neutron [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1036.011527] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1036.011789] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1036.011997] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Deleting the datastore file [datastore1] 5f248cf6-31a9-49ef-a8a8-218987664af7 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1036.012301] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e6b0126-1dda-4a5a-86aa-043a0e76a97f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.022037] env[61728]: DEBUG oslo_vmware.api [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Waiting for the task: (returnval){ [ 1036.022037] env[61728]: value = "task-464753" [ 1036.022037] env[61728]: _type = "Task" [ 1036.022037] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.030837] env[61728]: DEBUG oslo_vmware.api [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464753, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.048881] env[61728]: DEBUG nova.compute.manager [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1036.080318] env[61728]: DEBUG nova.virt.hardware [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1036.080614] env[61728]: DEBUG nova.virt.hardware [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1036.080807] env[61728]: DEBUG nova.virt.hardware [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1036.081171] env[61728]: DEBUG nova.virt.hardware [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1036.081385] env[61728]: DEBUG nova.virt.hardware [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1036.081623] env[61728]: DEBUG nova.virt.hardware [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1036.081962] env[61728]: DEBUG nova.virt.hardware [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1036.082283] env[61728]: DEBUG nova.virt.hardware [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1036.082583] env[61728]: DEBUG nova.virt.hardware [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1036.083225] env[61728]: DEBUG nova.virt.hardware [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1036.083225] env[61728]: DEBUG nova.virt.hardware [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1036.084725] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81abe31-91c3-49c0-9d3f-a290e1e0a729 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.093990] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09296654-bb5c-41ea-a904-ef2789f0f40e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.181572] env[61728]: DEBUG oslo_concurrency.lockutils [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Releasing lock "refresh_cache-705c8838-f605-46fa-b036-ec212f673f66" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.388191] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52791fb2-4486-3c68-9c15-3f05d2838e29, 'name': SearchDatastore_Task, 'duration_secs': 0.024572} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.392016] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.392016] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1036.392016] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.392016] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.392016] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1036.392016] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b0a5c80d-451b-4697-982a-5189d06af211 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.400860] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1036.400860] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1036.401343] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e745808-5f86-470a-bb8c-9f7a588fab12 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.408427] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1036.408427] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5201873c-8ad7-6ebc-db4c-142c05460b6c" [ 1036.408427] env[61728]: _type = "Task" [ 1036.408427] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.418221] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5201873c-8ad7-6ebc-db4c-142c05460b6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.426523] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.498786] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fb4fc573-ce99-4c75-8d64-13725163f2da tempest-InstanceActionsNegativeTestJSON-971437931 tempest-InstanceActionsNegativeTestJSON-971437931-project-member] Lock "9efdee89-4c88-4af3-af04-5281e8f6bc25" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.528s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.527939] env[61728]: DEBUG nova.network.neutron [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1036.534903] env[61728]: DEBUG oslo_vmware.api [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Task: {'id': task-464753, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.257834} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.537342] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.537629] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1036.537711] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1036.537887] env[61728]: INFO nova.compute.manager [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1036.538149] env[61728]: DEBUG oslo.service.loopingcall [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.538827] env[61728]: DEBUG nova.compute.manager [-] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1036.538827] env[61728]: DEBUG nova.network.neutron [-] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1036.690533] env[61728]: DEBUG oslo_concurrency.lockutils [None req-022b635c-379f-41ad-9b68-ef134e0d59fa tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "interface-705c8838-f605-46fa-b036-ec212f673f66-d0b341ea-4a0b-46ff-89f2-f96d212d04d9" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.842s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.714356] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d93acf-934f-4d2e-bf07-2d7c481589ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.722999] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b81da8-043c-4c38-8323-c6aa04cbbdb6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.757265] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bff2d1c-b8fe-4d25-95d4-1df1e0fd5256 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.765770] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-263e13b2-b59f-40fd-86e2-2323333d17ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.780366] env[61728]: DEBUG nova.compute.provider_tree [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.797619] env[61728]: DEBUG nova.network.neutron [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating instance_info_cache with network_info: [{"id": "e7f1c36b-89f0-4858-96a0-3c1875c4bb29", "address": "fa:16:3e:c0:51:65", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7f1c36b-89", "ovs_interfaceid": "e7f1c36b-89f0-4858-96a0-3c1875c4bb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.817480] env[61728]: DEBUG nova.compute.manager [req-7453e30c-3b4e-4aa4-90a5-21d5747157d5 req-3ca6b78f-32c1-4ca1-84de-53cea3e937d4 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Received event network-vif-deleted-ec0ae08d-62b0-4506-bc0b-65e9152dd5e8 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1036.817480] env[61728]: INFO nova.compute.manager [req-7453e30c-3b4e-4aa4-90a5-21d5747157d5 req-3ca6b78f-32c1-4ca1-84de-53cea3e937d4 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Neutron deleted interface ec0ae08d-62b0-4506-bc0b-65e9152dd5e8; detaching it from the instance and deleting it from the info cache [ 1036.817906] env[61728]: DEBUG nova.network.neutron [req-7453e30c-3b4e-4aa4-90a5-21d5747157d5 req-3ca6b78f-32c1-4ca1-84de-53cea3e937d4 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.847340] env[61728]: DEBUG nova.compute.manager [req-8d9abf38-d4a4-4143-9ca5-1d2f7c340fbe req-18dcc0cb-0352-4c75-8096-6153c61c10e1 service nova] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Received event network-vif-deleted-adc8e830-5757-48a7-8060-af7d189243df {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1036.847463] env[61728]: INFO nova.compute.manager [req-8d9abf38-d4a4-4143-9ca5-1d2f7c340fbe req-18dcc0cb-0352-4c75-8096-6153c61c10e1 service nova] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Neutron deleted interface adc8e830-5757-48a7-8060-af7d189243df; detaching it from the instance and deleting it from the info cache [ 1036.847753] env[61728]: DEBUG nova.network.neutron [req-8d9abf38-d4a4-4143-9ca5-1d2f7c340fbe req-18dcc0cb-0352-4c75-8096-6153c61c10e1 service nova] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.891212] env[61728]: DEBUG nova.network.neutron [-] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.926830] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5201873c-8ad7-6ebc-db4c-142c05460b6c, 'name': SearchDatastore_Task, 'duration_secs': 0.042688} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.926830] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-592e2cbc-5b22-45f3-adf4-30d001370b13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.935382] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1036.935382] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52cb6762-2cb7-8b5a-42ee-2cff5303c36c" [ 1036.935382] env[61728]: _type = "Task" [ 1036.935382] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.950372] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52cb6762-2cb7-8b5a-42ee-2cff5303c36c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.070867] env[61728]: DEBUG nova.network.neutron [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Successfully updated port: 63b2bf95-c628-4c62-bdbd-051192e13a2f {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1037.285043] env[61728]: DEBUG nova.scheduler.client.report [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1037.303740] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.303740] env[61728]: DEBUG nova.compute.manager [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Instance network_info: |[{"id": "e7f1c36b-89f0-4858-96a0-3c1875c4bb29", "address": "fa:16:3e:c0:51:65", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7f1c36b-89", "ovs_interfaceid": "e7f1c36b-89f0-4858-96a0-3c1875c4bb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1037.303740] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:51:65', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd733acc2-07d0-479e-918c-ec8a21925389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e7f1c36b-89f0-4858-96a0-3c1875c4bb29', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1037.311022] env[61728]: DEBUG oslo.service.loopingcall [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1037.312081] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1037.312324] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b70659b-763d-49f6-8e47-3328eab94271 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.326620] env[61728]: DEBUG nova.network.neutron [-] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.332340] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-93f4f4ee-4c9d-4ad2-96b9-1a5b1b4c057c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.337224] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1037.337224] env[61728]: value = "task-464754" [ 1037.337224] env[61728]: _type = "Task" [ 1037.337224] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.343848] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef0412a-6b64-41c1-ad46-d2b1c103c43d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.355158] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6667f3c6-c687-43ab-8fd3-1d3c7157646c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.361028] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464754, 'name': CreateVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.369387] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-327398b4-ffff-42a2-a03e-79a667f9f413 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.389940] env[61728]: DEBUG nova.compute.manager [req-7453e30c-3b4e-4aa4-90a5-21d5747157d5 req-3ca6b78f-32c1-4ca1-84de-53cea3e937d4 service nova] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Detach interface failed, port_id=ec0ae08d-62b0-4506-bc0b-65e9152dd5e8, reason: Instance 705c8838-f605-46fa-b036-ec212f673f66 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1037.397744] env[61728]: INFO nova.compute.manager [-] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Took 1.65 seconds to deallocate network for instance. [ 1037.424290] env[61728]: DEBUG nova.compute.manager [req-8d9abf38-d4a4-4143-9ca5-1d2f7c340fbe req-18dcc0cb-0352-4c75-8096-6153c61c10e1 service nova] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Detach interface failed, port_id=adc8e830-5757-48a7-8060-af7d189243df, reason: Instance 5f248cf6-31a9-49ef-a8a8-218987664af7 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1037.447138] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52cb6762-2cb7-8b5a-42ee-2cff5303c36c, 'name': SearchDatastore_Task, 'duration_secs': 0.011983} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.447454] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.447771] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] d55beb02-00fc-4df6-a239-9e5d776a7c18/d55beb02-00fc-4df6-a239-9e5d776a7c18.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1037.448057] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a25ffe3d-bcaf-44e1-ba6d-2f0133e3196f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.454733] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1037.454733] env[61728]: value = "task-464755" [ 1037.454733] env[61728]: _type = "Task" [ 1037.454733] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.462837] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464755, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.573175] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "refresh_cache-41329d0d-3838-4d1b-a4e6-8df4508eea76" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.573330] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "refresh_cache-41329d0d-3838-4d1b-a4e6-8df4508eea76" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.573552] env[61728]: DEBUG nova.network.neutron [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1037.788656] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.822s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.792076] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.386s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.793609] env[61728]: INFO nova.compute.claims [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1037.814533] env[61728]: INFO nova.scheduler.client.report [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleted allocations for instance bb7b38d5-53c7-4c05-96f8-9121171d3c46 [ 1037.829863] env[61728]: INFO nova.compute.manager [-] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Took 1.29 seconds to deallocate network for instance. [ 1037.851637] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464754, 'name': CreateVM_Task, 'duration_secs': 0.398982} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.851934] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1037.852718] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.852953] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.853339] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1037.853649] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-076cc772-1d05-4b04-8f3f-0e1014804a85 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.860241] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1037.860241] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f13c8a-0ea5-7499-c8d2-ac55db3667da" [ 1037.860241] env[61728]: _type = "Task" [ 1037.860241] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.870570] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f13c8a-0ea5-7499-c8d2-ac55db3667da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.925947] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.965523] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464755, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.120328] env[61728]: DEBUG nova.network.neutron [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1038.334172] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5c9f38e0-39a0-47c0-a2ea-62b443b87507 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "bb7b38d5-53c7-4c05-96f8-9121171d3c46" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.643s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.335569] env[61728]: DEBUG nova.network.neutron [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Updating instance_info_cache with network_info: [{"id": "63b2bf95-c628-4c62-bdbd-051192e13a2f", "address": "fa:16:3e:fc:4c:88", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63b2bf95-c6", "ovs_interfaceid": "63b2bf95-c628-4c62-bdbd-051192e13a2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.337381] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.373126] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f13c8a-0ea5-7499-c8d2-ac55db3667da, 'name': SearchDatastore_Task, 'duration_secs': 0.057128} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.373505] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.373789] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1038.374200] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.374359] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.374962] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1038.375164] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1874ab72-999e-4f8f-8de6-27d0ac7823a5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.384589] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1038.385142] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1038.385727] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-547e5ef5-1d88-4dda-b83b-866915e2d7f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.393077] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1038.393077] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52385d13-f1b8-4f80-e87b-0596be326610" [ 1038.393077] env[61728]: _type = "Task" [ 1038.393077] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.402118] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52385d13-f1b8-4f80-e87b-0596be326610, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.467170] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464755, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.521435] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "b4241d6f-1d70-4f61-a322-caff7c4e6f64" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.521693] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "b4241d6f-1d70-4f61-a322-caff7c4e6f64" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.521963] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "b4241d6f-1d70-4f61-a322-caff7c4e6f64-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.522177] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "b4241d6f-1d70-4f61-a322-caff7c4e6f64-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.522359] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "b4241d6f-1d70-4f61-a322-caff7c4e6f64-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.524812] env[61728]: INFO nova.compute.manager [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Terminating instance [ 1038.526748] env[61728]: DEBUG nova.compute.manager [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1038.526951] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1038.527796] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f0f678-da0d-4c73-9095-af2477f25746 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.535722] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1038.535960] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50480d80-cd2c-477c-b3c5-c591a420bdc2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.542558] env[61728]: DEBUG oslo_vmware.api [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1038.542558] env[61728]: value = "task-464756" [ 1038.542558] env[61728]: _type = "Task" [ 1038.542558] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.550696] env[61728]: DEBUG oslo_vmware.api [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464756, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.842023] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "refresh_cache-41329d0d-3838-4d1b-a4e6-8df4508eea76" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.842023] env[61728]: DEBUG nova.compute.manager [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Instance network_info: |[{"id": "63b2bf95-c628-4c62-bdbd-051192e13a2f", "address": "fa:16:3e:fc:4c:88", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63b2bf95-c6", "ovs_interfaceid": "63b2bf95-c628-4c62-bdbd-051192e13a2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1038.842023] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:4c:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9732690c-bdcf-4e6f-9a32-42c196333eb8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '63b2bf95-c628-4c62-bdbd-051192e13a2f', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1038.849537] env[61728]: DEBUG oslo.service.loopingcall [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1038.853520] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1038.854781] env[61728]: DEBUG nova.compute.manager [req-ae540f3c-2f24-4bfc-b6d2-7d5a706a5374 req-efb3bed5-a74f-412c-bf6b-8e5dbbfb5f64 service nova] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Received event network-vif-plugged-63b2bf95-c628-4c62-bdbd-051192e13a2f {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1038.855134] env[61728]: DEBUG oslo_concurrency.lockutils [req-ae540f3c-2f24-4bfc-b6d2-7d5a706a5374 req-efb3bed5-a74f-412c-bf6b-8e5dbbfb5f64 service nova] Acquiring lock "41329d0d-3838-4d1b-a4e6-8df4508eea76-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.855454] env[61728]: DEBUG oslo_concurrency.lockutils [req-ae540f3c-2f24-4bfc-b6d2-7d5a706a5374 req-efb3bed5-a74f-412c-bf6b-8e5dbbfb5f64 service nova] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.855739] env[61728]: DEBUG oslo_concurrency.lockutils [req-ae540f3c-2f24-4bfc-b6d2-7d5a706a5374 req-efb3bed5-a74f-412c-bf6b-8e5dbbfb5f64 service nova] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.856030] env[61728]: DEBUG nova.compute.manager [req-ae540f3c-2f24-4bfc-b6d2-7d5a706a5374 req-efb3bed5-a74f-412c-bf6b-8e5dbbfb5f64 service nova] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] No waiting events found dispatching network-vif-plugged-63b2bf95-c628-4c62-bdbd-051192e13a2f {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1038.856330] env[61728]: WARNING nova.compute.manager [req-ae540f3c-2f24-4bfc-b6d2-7d5a706a5374 req-efb3bed5-a74f-412c-bf6b-8e5dbbfb5f64 service nova] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Received unexpected event network-vif-plugged-63b2bf95-c628-4c62-bdbd-051192e13a2f for instance with vm_state building and task_state spawning. [ 1038.856609] env[61728]: DEBUG nova.compute.manager [req-ae540f3c-2f24-4bfc-b6d2-7d5a706a5374 req-efb3bed5-a74f-412c-bf6b-8e5dbbfb5f64 service nova] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Received event network-changed-63b2bf95-c628-4c62-bdbd-051192e13a2f {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1038.856888] env[61728]: DEBUG nova.compute.manager [req-ae540f3c-2f24-4bfc-b6d2-7d5a706a5374 req-efb3bed5-a74f-412c-bf6b-8e5dbbfb5f64 service nova] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Refreshing instance network info cache due to event network-changed-63b2bf95-c628-4c62-bdbd-051192e13a2f. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1038.857234] env[61728]: DEBUG oslo_concurrency.lockutils [req-ae540f3c-2f24-4bfc-b6d2-7d5a706a5374 req-efb3bed5-a74f-412c-bf6b-8e5dbbfb5f64 service nova] Acquiring lock "refresh_cache-41329d0d-3838-4d1b-a4e6-8df4508eea76" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.857493] env[61728]: DEBUG oslo_concurrency.lockutils [req-ae540f3c-2f24-4bfc-b6d2-7d5a706a5374 req-efb3bed5-a74f-412c-bf6b-8e5dbbfb5f64 service nova] Acquired lock "refresh_cache-41329d0d-3838-4d1b-a4e6-8df4508eea76" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.857780] env[61728]: DEBUG nova.network.neutron [req-ae540f3c-2f24-4bfc-b6d2-7d5a706a5374 req-efb3bed5-a74f-412c-bf6b-8e5dbbfb5f64 service nova] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Refreshing network info cache for port 63b2bf95-c628-4c62-bdbd-051192e13a2f {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1038.859935] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3a38501-ed8f-4d4a-b435-703b16962f25 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.887193] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1038.887193] env[61728]: value = "task-464757" [ 1038.887193] env[61728]: _type = "Task" [ 1038.887193] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.899195] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464757, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.904894] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52385d13-f1b8-4f80-e87b-0596be326610, 'name': SearchDatastore_Task, 'duration_secs': 0.00973} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.908057] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc47b8c9-37fb-42c5-bcf6-c39981793929 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.913049] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1038.913049] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d286c3-62a6-3cc4-1f9d-dcec4ccd4556" [ 1038.913049] env[61728]: _type = "Task" [ 1038.913049] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.922243] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d286c3-62a6-3cc4-1f9d-dcec4ccd4556, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.971800] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464755, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.498638} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.972284] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] d55beb02-00fc-4df6-a239-9e5d776a7c18/d55beb02-00fc-4df6-a239-9e5d776a7c18.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1038.972604] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1038.972969] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c4c935a-6117-4e03-8a36-95e1674aad82 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.984238] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1038.984238] env[61728]: value = "task-464758" [ 1038.984238] env[61728]: _type = "Task" [ 1038.984238] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.994121] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464758, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.055171] env[61728]: DEBUG oslo_vmware.api [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464756, 'name': PowerOffVM_Task, 'duration_secs': 0.19976} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.056641] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1039.056906] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1039.057994] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7372861b-481e-4275-a725-0f08289394fb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.061645] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8997b528-d8ca-488f-aca5-32b953214864 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.069047] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ecef9c0-2006-4524-903d-408b3992c4b5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.104877] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f10d685-3e75-4de1-a3ff-5c447a882927 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.112810] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c46118-7235-4f49-9179-658637c80b6f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.131796] env[61728]: DEBUG nova.compute.provider_tree [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.134230] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1039.134491] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1039.134712] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleting the datastore file [datastore1] b4241d6f-1d70-4f61-a322-caff7c4e6f64 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1039.135241] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8f233e2-c0c7-4437-b702-54f04e459de5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.141993] env[61728]: DEBUG oslo_vmware.api [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1039.141993] env[61728]: value = "task-464760" [ 1039.141993] env[61728]: _type = "Task" [ 1039.141993] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.151456] env[61728]: DEBUG oslo_vmware.api [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464760, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.396586] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464757, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.422630] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d286c3-62a6-3cc4-1f9d-dcec4ccd4556, 'name': SearchDatastore_Task, 'duration_secs': 0.0088} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.422839] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.423127] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 6c154517-68e8-4e70-86eb-cf53f6f812d8/6c154517-68e8-4e70-86eb-cf53f6f812d8.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1039.423386] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c5828a02-b677-497c-a57f-1b304e73e029 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.428948] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1039.428948] env[61728]: value = "task-464761" [ 1039.428948] env[61728]: _type = "Task" [ 1039.428948] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.438155] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464761, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.494066] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464758, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107408} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.494485] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1039.497564] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef8173d-b843-4102-948d-f14cdfee6f2f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.522412] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] d55beb02-00fc-4df6-a239-9e5d776a7c18/d55beb02-00fc-4df6-a239-9e5d776a7c18.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1039.522730] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88401219-0d2e-4fb7-92d7-ca3b55512c0a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.544494] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1039.544494] env[61728]: value = "task-464762" [ 1039.544494] env[61728]: _type = "Task" [ 1039.544494] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.552540] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464762, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.573781] env[61728]: DEBUG nova.network.neutron [req-ae540f3c-2f24-4bfc-b6d2-7d5a706a5374 req-efb3bed5-a74f-412c-bf6b-8e5dbbfb5f64 service nova] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Updated VIF entry in instance network info cache for port 63b2bf95-c628-4c62-bdbd-051192e13a2f. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1039.574271] env[61728]: DEBUG nova.network.neutron [req-ae540f3c-2f24-4bfc-b6d2-7d5a706a5374 req-efb3bed5-a74f-412c-bf6b-8e5dbbfb5f64 service nova] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Updating instance_info_cache with network_info: [{"id": "63b2bf95-c628-4c62-bdbd-051192e13a2f", "address": "fa:16:3e:fc:4c:88", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63b2bf95-c6", "ovs_interfaceid": "63b2bf95-c628-4c62-bdbd-051192e13a2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.636945] env[61728]: DEBUG nova.scheduler.client.report [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.653799] env[61728]: DEBUG oslo_vmware.api [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464760, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.275486} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.654157] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.654363] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1039.654552] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1039.654834] env[61728]: INFO nova.compute.manager [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1039.655009] env[61728]: DEBUG oslo.service.loopingcall [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.655212] env[61728]: DEBUG nova.compute.manager [-] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1039.655322] env[61728]: DEBUG nova.network.neutron [-] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1039.898771] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464757, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.939575] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464761, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501029} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.939879] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 6c154517-68e8-4e70-86eb-cf53f6f812d8/6c154517-68e8-4e70-86eb-cf53f6f812d8.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1039.940116] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1039.940375] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dea5dd98-3907-4a5c-973a-780630bb530c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.947105] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1039.947105] env[61728]: value = "task-464763" [ 1039.947105] env[61728]: _type = "Task" [ 1039.947105] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.954961] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464763, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.054929] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464762, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.077275] env[61728]: DEBUG oslo_concurrency.lockutils [req-ae540f3c-2f24-4bfc-b6d2-7d5a706a5374 req-efb3bed5-a74f-412c-bf6b-8e5dbbfb5f64 service nova] Releasing lock "refresh_cache-41329d0d-3838-4d1b-a4e6-8df4508eea76" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.143534] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.144201] env[61728]: DEBUG nova.compute.manager [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1040.147027] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.243s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.148448] env[61728]: INFO nova.compute.claims [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1040.398393] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464757, 'name': CreateVM_Task, 'duration_secs': 1.332313} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.398689] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1040.399225] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.399404] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.399779] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1040.400047] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6293e416-0816-4b54-bd11-ea80101afb26 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.404145] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1040.404145] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522a7fcd-cf31-f1f1-9506-6e18e6c71351" [ 1040.404145] env[61728]: _type = "Task" [ 1040.404145] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.411807] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522a7fcd-cf31-f1f1-9506-6e18e6c71351, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.421182] env[61728]: DEBUG nova.network.neutron [-] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.455099] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464763, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068486} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.455360] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1040.456085] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9e0096-1ee6-422f-88fd-d7e9fedcfbc7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.478009] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 6c154517-68e8-4e70-86eb-cf53f6f812d8/6c154517-68e8-4e70-86eb-cf53f6f812d8.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.478250] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a2f1608-810e-4671-9d9b-cfe2a55efbb1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.496054] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1040.496054] env[61728]: value = "task-464764" [ 1040.496054] env[61728]: _type = "Task" [ 1040.496054] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.503496] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464764, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.555059] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464762, 'name': ReconfigVM_Task, 'duration_secs': 0.52723} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.555617] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Reconfigured VM instance instance-00000069 to attach disk [datastore1] d55beb02-00fc-4df6-a239-9e5d776a7c18/d55beb02-00fc-4df6-a239-9e5d776a7c18.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.556450] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d77d43d1-6a0f-48e1-9687-4e53a66aa2d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.562440] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1040.562440] env[61728]: value = "task-464765" [ 1040.562440] env[61728]: _type = "Task" [ 1040.562440] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.570549] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464765, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.652755] env[61728]: DEBUG nova.compute.utils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1040.657778] env[61728]: DEBUG nova.compute.manager [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1040.657778] env[61728]: DEBUG nova.network.neutron [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1040.709625] env[61728]: DEBUG nova.policy [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31a2286aad9a40da9b16353541a2780f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49c64edaa37f41d38aba7bd5b1d0b47f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1040.871135] env[61728]: DEBUG nova.compute.manager [req-a97a9431-0aeb-4588-95e4-33db5b580b4b req-2974c3d6-28af-4c4a-a816-640c12dfae75 service nova] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Received event network-vif-deleted-81ffb8f2-c59b-4b0e-894a-d8633440d6eb {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1040.915064] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522a7fcd-cf31-f1f1-9506-6e18e6c71351, 'name': SearchDatastore_Task, 'duration_secs': 0.009012} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.915390] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.915632] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1040.915871] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.916036] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.916231] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1040.916683] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-14e2def0-beb7-454f-a8c2-f23b8e0fdc83 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.924817] env[61728]: INFO nova.compute.manager [-] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Took 1.27 seconds to deallocate network for instance. [ 1040.925130] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1040.925309] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1040.928057] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76fda787-7920-4243-9bd7-fc85b4948030 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.936568] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1040.936568] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526f1762-e1e1-cea9-3213-c55d3398f0bc" [ 1040.936568] env[61728]: _type = "Task" [ 1040.936568] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.945261] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526f1762-e1e1-cea9-3213-c55d3398f0bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.002049] env[61728]: DEBUG nova.network.neutron [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Successfully created port: ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1041.011953] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464764, 'name': ReconfigVM_Task, 'duration_secs': 0.275852} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.011953] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 6c154517-68e8-4e70-86eb-cf53f6f812d8/6c154517-68e8-4e70-86eb-cf53f6f812d8.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1041.011953] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ec968bc1-fb19-4e39-ad1b-e34dc3e05a98 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.017085] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1041.017085] env[61728]: value = "task-464766" [ 1041.017085] env[61728]: _type = "Task" [ 1041.017085] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.024773] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464766, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.071796] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464765, 'name': Rename_Task, 'duration_secs': 0.202612} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.072106] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1041.072703] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12a84e7f-404b-45a9-9701-d5915a055670 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.078023] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1041.078023] env[61728]: value = "task-464767" [ 1041.078023] env[61728]: _type = "Task" [ 1041.078023] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.085591] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464767, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.159186] env[61728]: DEBUG nova.compute.manager [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1041.405684] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e9d4f0-321d-429a-9829-0e6a967de326 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.414959] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ed0faf6-bf45-4a00-800b-b46daa8b21be {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.444100] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.447797] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c04b27f-0bd4-4565-bae3-6dc57200731b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.458377] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e23575-6a51-4fd3-83f3-8461158e9372 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.462475] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526f1762-e1e1-cea9-3213-c55d3398f0bc, 'name': SearchDatastore_Task, 'duration_secs': 0.014169} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.463577] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54c3374c-bb3e-468e-bdc7-a2fd77c3d15e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.475587] env[61728]: DEBUG nova.compute.provider_tree [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.479515] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1041.479515] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52091c24-2e4a-9f4a-2d47-de0bbc499dbe" [ 1041.479515] env[61728]: _type = "Task" [ 1041.479515] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.487045] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52091c24-2e4a-9f4a-2d47-de0bbc499dbe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.527359] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464766, 'name': Rename_Task, 'duration_secs': 0.148432} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.527649] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1041.527903] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a5f768d-324a-4e01-b178-79157987cf32 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.534249] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1041.534249] env[61728]: value = "task-464768" [ 1041.534249] env[61728]: _type = "Task" [ 1041.534249] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.541569] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464768, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.589267] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464767, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.978917] env[61728]: DEBUG nova.scheduler.client.report [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.992929] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52091c24-2e4a-9f4a-2d47-de0bbc499dbe, 'name': SearchDatastore_Task, 'duration_secs': 0.009503} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.993257] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.993525] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 41329d0d-3838-4d1b-a4e6-8df4508eea76/41329d0d-3838-4d1b-a4e6-8df4508eea76.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1041.993837] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a57e946c-1d9f-4bd0-bb3b-5d1610261d51 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.002389] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1042.002389] env[61728]: value = "task-464769" [ 1042.002389] env[61728]: _type = "Task" [ 1042.002389] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.011071] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464769, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.044041] env[61728]: DEBUG oslo_vmware.api [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464768, 'name': PowerOnVM_Task, 'duration_secs': 0.502437} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.044328] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1042.044542] env[61728]: INFO nova.compute.manager [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Took 8.39 seconds to spawn the instance on the hypervisor. [ 1042.044755] env[61728]: DEBUG nova.compute.manager [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.045572] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e5d028-0e79-422b-a964-327e67024d7b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.088451] env[61728]: DEBUG oslo_vmware.api [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464767, 'name': PowerOnVM_Task, 'duration_secs': 0.526832} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.088714] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1042.088922] env[61728]: INFO nova.compute.manager [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Took 10.87 seconds to spawn the instance on the hypervisor. [ 1042.089325] env[61728]: DEBUG nova.compute.manager [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.090753] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d23c5d1-3ad8-4184-9112-10d000cad524 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.172663] env[61728]: DEBUG nova.compute.manager [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1042.201327] env[61728]: DEBUG nova.virt.hardware [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1042.201597] env[61728]: DEBUG nova.virt.hardware [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1042.201828] env[61728]: DEBUG nova.virt.hardware [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1042.202070] env[61728]: DEBUG nova.virt.hardware [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1042.202242] env[61728]: DEBUG nova.virt.hardware [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1042.202403] env[61728]: DEBUG nova.virt.hardware [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1042.202621] env[61728]: DEBUG nova.virt.hardware [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1042.202837] env[61728]: DEBUG nova.virt.hardware [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1042.203065] env[61728]: DEBUG nova.virt.hardware [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1042.203251] env[61728]: DEBUG nova.virt.hardware [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1042.203437] env[61728]: DEBUG nova.virt.hardware [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1042.204303] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32fa5613-481a-4c61-a98c-0ee57bed28ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.211940] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6072079-e47d-4493-a489-a505fd06d55f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.462698] env[61728]: DEBUG nova.network.neutron [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Successfully updated port: ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1042.487709] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.341s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.488298] env[61728]: DEBUG nova.compute.manager [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1042.491425] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.065s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.491661] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.491872] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61728) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1042.492206] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.566s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.492447] env[61728]: DEBUG nova.objects.instance [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lazy-loading 'resources' on Instance uuid 705c8838-f605-46fa-b036-ec212f673f66 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.495046] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41374b1-8a18-4a1a-9f8b-f0363b37555f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.507153] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422f6e31-bf8b-43af-b7b6-263a82c7895e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.529572] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e627ef-4ad3-4c1a-abe2-75ce0f8dabae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.533539] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464769, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.540643] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c927c79-0600-44ea-8a32-d186907ac3e5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.583752] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180230MB free_disk=114GB free_vcpus=48 pci_devices=None {{(pid=61728) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1042.584316] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.585213] env[61728]: INFO nova.compute.manager [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Took 24.78 seconds to build instance. [ 1042.608643] env[61728]: INFO nova.compute.manager [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Took 25.27 seconds to build instance. [ 1042.896250] env[61728]: DEBUG nova.compute.manager [req-a6c6d8b9-0332-4347-b9c0-9950d3befe06 req-1f07ffa8-0603-44fe-870f-6fd34ef69d55 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Received event network-vif-plugged-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1042.896486] env[61728]: DEBUG oslo_concurrency.lockutils [req-a6c6d8b9-0332-4347-b9c0-9950d3befe06 req-1f07ffa8-0603-44fe-870f-6fd34ef69d55 service nova] Acquiring lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.896723] env[61728]: DEBUG oslo_concurrency.lockutils [req-a6c6d8b9-0332-4347-b9c0-9950d3befe06 req-1f07ffa8-0603-44fe-870f-6fd34ef69d55 service nova] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.896892] env[61728]: DEBUG oslo_concurrency.lockutils [req-a6c6d8b9-0332-4347-b9c0-9950d3befe06 req-1f07ffa8-0603-44fe-870f-6fd34ef69d55 service nova] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.897081] env[61728]: DEBUG nova.compute.manager [req-a6c6d8b9-0332-4347-b9c0-9950d3befe06 req-1f07ffa8-0603-44fe-870f-6fd34ef69d55 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] No waiting events found dispatching network-vif-plugged-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1042.897257] env[61728]: WARNING nova.compute.manager [req-a6c6d8b9-0332-4347-b9c0-9950d3befe06 req-1f07ffa8-0603-44fe-870f-6fd34ef69d55 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Received unexpected event network-vif-plugged-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb for instance with vm_state building and task_state spawning. [ 1042.897422] env[61728]: DEBUG nova.compute.manager [req-a6c6d8b9-0332-4347-b9c0-9950d3befe06 req-1f07ffa8-0603-44fe-870f-6fd34ef69d55 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Received event network-changed-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1042.897594] env[61728]: DEBUG nova.compute.manager [req-a6c6d8b9-0332-4347-b9c0-9950d3befe06 req-1f07ffa8-0603-44fe-870f-6fd34ef69d55 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Refreshing instance network info cache due to event network-changed-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1042.898033] env[61728]: DEBUG oslo_concurrency.lockutils [req-a6c6d8b9-0332-4347-b9c0-9950d3befe06 req-1f07ffa8-0603-44fe-870f-6fd34ef69d55 service nova] Acquiring lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.899145] env[61728]: DEBUG oslo_concurrency.lockutils [req-a6c6d8b9-0332-4347-b9c0-9950d3befe06 req-1f07ffa8-0603-44fe-870f-6fd34ef69d55 service nova] Acquired lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.899145] env[61728]: DEBUG nova.network.neutron [req-a6c6d8b9-0332-4347-b9c0-9950d3befe06 req-1f07ffa8-0603-44fe-870f-6fd34ef69d55 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Refreshing network info cache for port ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1042.965604] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.995753] env[61728]: DEBUG nova.compute.utils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1042.997523] env[61728]: DEBUG nova.compute.manager [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1042.997678] env[61728]: DEBUG nova.network.neutron [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1043.014327] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464769, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.820552} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.015380] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 41329d0d-3838-4d1b-a4e6-8df4508eea76/41329d0d-3838-4d1b-a4e6-8df4508eea76.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1043.015546] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1043.015853] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-32cc25e2-9f63-444a-aa22-d2080c00ee19 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.025730] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1043.025730] env[61728]: value = "task-464770" [ 1043.025730] env[61728]: _type = "Task" [ 1043.025730] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.036727] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464770, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.045064] env[61728]: DEBUG nova.policy [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ee5c541201ef47c1b028f5be9f853e59', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd88c08b899c94efb8254e510bdf53d25', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1043.088133] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d8c4f599-9156-47e1-b033-3c7a74d99521 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "6c154517-68e8-4e70-86eb-cf53f6f812d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.293s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.114289] env[61728]: DEBUG oslo_concurrency.lockutils [None req-afa76cd7-da95-4890-8cd0-7a1ba28f25af tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.782s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.204978] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f56cea-e57e-47e4-841c-d3e40968f05c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.212998] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc975ba-ac86-4b12-8e11-5aadbabd8d38 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.244311] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbc04b8-7c50-450c-bb67-6352fd9c2fad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.252019] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d8c030-c3cc-4d48-8753-002e9a9efb79 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.265131] env[61728]: DEBUG nova.compute.provider_tree [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.445016] env[61728]: DEBUG nova.network.neutron [req-a6c6d8b9-0332-4347-b9c0-9950d3befe06 req-1f07ffa8-0603-44fe-870f-6fd34ef69d55 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1043.500898] env[61728]: DEBUG nova.compute.manager [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1043.536356] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464770, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073763} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.536648] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1043.540818] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40f5269-49b9-4703-8693-d88153262a83 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.564076] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 41329d0d-3838-4d1b-a4e6-8df4508eea76/41329d0d-3838-4d1b-a4e6-8df4508eea76.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1043.565728] env[61728]: DEBUG nova.network.neutron [req-a6c6d8b9-0332-4347-b9c0-9950d3befe06 req-1f07ffa8-0603-44fe-870f-6fd34ef69d55 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.566976] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef7ebbc0-4671-4776-99ad-26afb8400af5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.583278] env[61728]: DEBUG oslo_concurrency.lockutils [req-a6c6d8b9-0332-4347-b9c0-9950d3befe06 req-1f07ffa8-0603-44fe-870f-6fd34ef69d55 service nova] Releasing lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.584939] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.584939] env[61728]: DEBUG nova.network.neutron [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1043.590932] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1043.590932] env[61728]: value = "task-464771" [ 1043.590932] env[61728]: _type = "Task" [ 1043.590932] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.602733] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464771, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.603754] env[61728]: DEBUG nova.network.neutron [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Successfully created port: 56270ade-0651-4e82-8fdd-4bcc30bfd53c {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1043.768291] env[61728]: DEBUG nova.scheduler.client.report [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.946228] env[61728]: DEBUG nova.compute.manager [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Stashing vm_state: active {{(pid=61728) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1044.101667] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464771, 'name': ReconfigVM_Task, 'duration_secs': 0.38018} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.102215] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 41329d0d-3838-4d1b-a4e6-8df4508eea76/41329d0d-3838-4d1b-a4e6-8df4508eea76.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1044.102998] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ab5f14d-dcd2-4df7-a780-c5b7f3d76ada {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.109988] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1044.109988] env[61728]: value = "task-464772" [ 1044.109988] env[61728]: _type = "Task" [ 1044.109988] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.117833] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464772, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.118639] env[61728]: DEBUG nova.network.neutron [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1044.255491] env[61728]: DEBUG nova.network.neutron [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Updating instance_info_cache with network_info: [{"id": "ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb", "address": "fa:16:3e:36:bc:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea6ae87d-d3", "ovs_interfaceid": "ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.273448] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.781s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.276047] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.939s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.276853] env[61728]: DEBUG nova.objects.instance [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Lazy-loading 'resources' on Instance uuid 5f248cf6-31a9-49ef-a8a8-218987664af7 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.296495] env[61728]: INFO nova.scheduler.client.report [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Deleted allocations for instance 705c8838-f605-46fa-b036-ec212f673f66 [ 1044.463022] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.512779] env[61728]: DEBUG nova.compute.manager [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1044.540789] env[61728]: DEBUG nova.virt.hardware [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.541519] env[61728]: DEBUG nova.virt.hardware [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.541519] env[61728]: DEBUG nova.virt.hardware [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.541519] env[61728]: DEBUG nova.virt.hardware [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.541683] env[61728]: DEBUG nova.virt.hardware [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.541839] env[61728]: DEBUG nova.virt.hardware [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.542101] env[61728]: DEBUG nova.virt.hardware [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.542279] env[61728]: DEBUG nova.virt.hardware [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.542462] env[61728]: DEBUG nova.virt.hardware [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.542636] env[61728]: DEBUG nova.virt.hardware [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.542816] env[61728]: DEBUG nova.virt.hardware [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.543790] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4bcd7f-0512-49f1-bcb7-2fa0fd4c0115 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.552513] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740d44c8-3911-43ea-bd5f-5aea1408b92f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.623894] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464772, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.758796] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.759255] env[61728]: DEBUG nova.compute.manager [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Instance network_info: |[{"id": "ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb", "address": "fa:16:3e:36:bc:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea6ae87d-d3", "ovs_interfaceid": "ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1044.759802] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:bc:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e51ebca-e0f8-4b77-b155-4ff928eef130', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1044.768346] env[61728]: DEBUG oslo.service.loopingcall [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.768621] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1044.768918] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-440ccc66-06c9-4d25-9361-96d4d0bfe177 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.792359] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1044.792359] env[61728]: value = "task-464773" [ 1044.792359] env[61728]: _type = "Task" [ 1044.792359] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.804271] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464773, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.807631] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c4f9b598-1f63-45a4-983b-6ec389fdba41 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "705c8838-f605-46fa-b036-ec212f673f66" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.188s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.924786] env[61728]: DEBUG nova.compute.manager [req-c44c8b92-26a7-4ae7-a6d3-ef99549ff31b req-60811dc6-4cc6-40c5-a46c-f94cef1a02f7 service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Received event network-changed-4155b1a7-975f-404c-8374-47665c8046ca {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1044.925497] env[61728]: DEBUG nova.compute.manager [req-c44c8b92-26a7-4ae7-a6d3-ef99549ff31b req-60811dc6-4cc6-40c5-a46c-f94cef1a02f7 service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Refreshing instance network info cache due to event network-changed-4155b1a7-975f-404c-8374-47665c8046ca. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1044.925789] env[61728]: DEBUG oslo_concurrency.lockutils [req-c44c8b92-26a7-4ae7-a6d3-ef99549ff31b req-60811dc6-4cc6-40c5-a46c-f94cef1a02f7 service nova] Acquiring lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.926044] env[61728]: DEBUG oslo_concurrency.lockutils [req-c44c8b92-26a7-4ae7-a6d3-ef99549ff31b req-60811dc6-4cc6-40c5-a46c-f94cef1a02f7 service nova] Acquired lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.926240] env[61728]: DEBUG nova.network.neutron [req-c44c8b92-26a7-4ae7-a6d3-ef99549ff31b req-60811dc6-4cc6-40c5-a46c-f94cef1a02f7 service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Refreshing network info cache for port 4155b1a7-975f-404c-8374-47665c8046ca {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1045.019459] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-976e7670-e7e6-4f0c-a146-748c2d5c2c75 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.027792] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6303f584-61c3-48f8-8968-0fe5c217160d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.057565] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358dd43c-a18c-4ceb-97bd-3b8d921c2321 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.065611] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d0e521-d60b-4e3c-9453-aac5f03c1637 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.081629] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "a70d1c37-e792-4168-b7e1-9418b8cb7818" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.081929] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "a70d1c37-e792-4168-b7e1-9418b8cb7818" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.082166] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "a70d1c37-e792-4168-b7e1-9418b8cb7818-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.082361] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "a70d1c37-e792-4168-b7e1-9418b8cb7818-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.082534] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "a70d1c37-e792-4168-b7e1-9418b8cb7818-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.084451] env[61728]: DEBUG nova.compute.provider_tree [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.085734] env[61728]: INFO nova.compute.manager [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Terminating instance [ 1045.090852] env[61728]: DEBUG nova.compute.manager [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1045.091070] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1045.091883] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adead902-79fc-403a-aa8e-5feadffffa4f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.099446] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1045.099679] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99fd4680-d515-4238-a590-4becc0aa63ac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.106132] env[61728]: DEBUG oslo_vmware.api [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 1045.106132] env[61728]: value = "task-464774" [ 1045.106132] env[61728]: _type = "Task" [ 1045.106132] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.113747] env[61728]: DEBUG oslo_vmware.api [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464774, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.122027] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464772, 'name': Rename_Task, 'duration_secs': 0.915423} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.122027] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1045.122027] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-723850ef-0baa-46f1-aafc-1118fc37dae5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.129070] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1045.129070] env[61728]: value = "task-464775" [ 1045.129070] env[61728]: _type = "Task" [ 1045.129070] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.138535] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464775, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.303172] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464773, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.335263] env[61728]: DEBUG nova.network.neutron [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Successfully updated port: 56270ade-0651-4e82-8fdd-4bcc30bfd53c {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1045.588464] env[61728]: DEBUG nova.scheduler.client.report [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.622819] env[61728]: DEBUG oslo_vmware.api [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464774, 'name': PowerOffVM_Task, 'duration_secs': 0.268006} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.622819] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1045.623067] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1045.623339] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aeeba379-5ce3-4283-93b4-4c678ca9895c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.638477] env[61728]: DEBUG oslo_vmware.api [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464775, 'name': PowerOnVM_Task, 'duration_secs': 0.507096} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.638726] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1045.638936] env[61728]: INFO nova.compute.manager [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Took 9.59 seconds to spawn the instance on the hypervisor. [ 1045.639131] env[61728]: DEBUG nova.compute.manager [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1045.639867] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da354c4b-5eaf-47ba-84c2-e5713510161b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.649063] env[61728]: DEBUG nova.network.neutron [req-c44c8b92-26a7-4ae7-a6d3-ef99549ff31b req-60811dc6-4cc6-40c5-a46c-f94cef1a02f7 service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Updated VIF entry in instance network info cache for port 4155b1a7-975f-404c-8374-47665c8046ca. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1045.649460] env[61728]: DEBUG nova.network.neutron [req-c44c8b92-26a7-4ae7-a6d3-ef99549ff31b req-60811dc6-4cc6-40c5-a46c-f94cef1a02f7 service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Updating instance_info_cache with network_info: [{"id": "4155b1a7-975f-404c-8374-47665c8046ca", "address": "fa:16:3e:e4:b3:9b", "network": {"id": "72d11e2a-907e-419e-b8a4-53aa39f2566a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1755247352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833a646ea21a48c4951ed8cedebaa98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3836fb52-19c6-4e10-a0ca-f0bca73dc887", "external-id": "nsx-vlan-transportzone-964", "segmentation_id": 964, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4155b1a7-97", "ovs_interfaceid": "4155b1a7-975f-404c-8374-47665c8046ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.687027] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1045.687288] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1045.687482] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Deleting the datastore file [datastore1] a70d1c37-e792-4168-b7e1-9418b8cb7818 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1045.688402] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7bac2d1-a099-4793-80e4-9bc724ce4d00 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.696763] env[61728]: DEBUG oslo_vmware.api [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for the task: (returnval){ [ 1045.696763] env[61728]: value = "task-464777" [ 1045.696763] env[61728]: _type = "Task" [ 1045.696763] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.705023] env[61728]: DEBUG oslo_vmware.api [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464777, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.803473] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464773, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.837714] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquiring lock "refresh_cache-e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.837908] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquired lock "refresh_cache-e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.838984] env[61728]: DEBUG nova.network.neutron [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1046.093606] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.817s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.099680] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.652s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.099680] env[61728]: DEBUG nova.objects.instance [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lazy-loading 'resources' on Instance uuid b4241d6f-1d70-4f61-a322-caff7c4e6f64 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.122626] env[61728]: INFO nova.scheduler.client.report [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Deleted allocations for instance 5f248cf6-31a9-49ef-a8a8-218987664af7 [ 1046.157052] env[61728]: DEBUG oslo_concurrency.lockutils [req-c44c8b92-26a7-4ae7-a6d3-ef99549ff31b req-60811dc6-4cc6-40c5-a46c-f94cef1a02f7 service nova] Releasing lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.158624] env[61728]: INFO nova.compute.manager [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Took 16.41 seconds to build instance. [ 1046.174392] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "267880ff-0fac-4763-82be-ef8587ee7628" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.174641] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "267880ff-0fac-4763-82be-ef8587ee7628" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.207133] env[61728]: DEBUG oslo_vmware.api [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Task: {'id': task-464777, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.421457} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.207447] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1046.207648] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1046.207830] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1046.208074] env[61728]: INFO nova.compute.manager [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1046.208365] env[61728]: DEBUG oslo.service.loopingcall [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.208565] env[61728]: DEBUG nova.compute.manager [-] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1046.208662] env[61728]: DEBUG nova.network.neutron [-] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1046.304905] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464773, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.368216] env[61728]: DEBUG nova.network.neutron [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1046.517338] env[61728]: DEBUG nova.network.neutron [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Updating instance_info_cache with network_info: [{"id": "56270ade-0651-4e82-8fdd-4bcc30bfd53c", "address": "fa:16:3e:35:57:51", "network": {"id": "11128295-2780-4c0a-81cf-946216b8d623", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-492544146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d88c08b899c94efb8254e510bdf53d25", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56270ade-06", "ovs_interfaceid": "56270ade-0651-4e82-8fdd-4bcc30bfd53c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.632306] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fadaf688-3c8c-4fc9-bc6b-259d1f744c6e tempest-ServerAddressesNegativeTestJSON-1984082900 tempest-ServerAddressesNegativeTestJSON-1984082900-project-member] Lock "5f248cf6-31a9-49ef-a8a8-218987664af7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.210s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.659346] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d075dce-670d-422b-83d1-28061c949e9b tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.915s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.679933] env[61728]: DEBUG nova.compute.utils [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1046.806526] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464773, 'name': CreateVM_Task, 'duration_secs': 1.55864} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.806775] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1046.807402] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.807599] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.807937] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1046.808296] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49ad63d2-0a39-40cb-b5d5-7da5675105cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.811592] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e950ae2-f008-4a99-94fa-6f26f7509c5f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.818224] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1046.818224] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e042fc-0313-fd56-f015-e5322f7ffa6a" [ 1046.818224] env[61728]: _type = "Task" [ 1046.818224] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.824808] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-124f12af-ecd5-443e-9c5a-bcb018461017 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.834306] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e042fc-0313-fd56-f015-e5322f7ffa6a, 'name': SearchDatastore_Task, 'duration_secs': 0.009821} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.864381] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.864690] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1046.864942] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.865112] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.865304] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1046.865816] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29204188-6fbb-44d8-9ad0-7642ecb4e62b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.868236] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7d6161-e5d4-42c0-9a82-38b4159dd457 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.876606] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577bd23f-c848-46f0-9e6a-ca8df0ab87ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.881412] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1046.881601] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1046.882625] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f1159ca-5158-461f-9c99-df3fcf59a59a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.893522] env[61728]: DEBUG nova.compute.provider_tree [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.898032] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1046.898032] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520cd4e5-3e84-52ab-2ff7-f1714f9713f7" [ 1046.898032] env[61728]: _type = "Task" [ 1046.898032] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.906024] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520cd4e5-3e84-52ab-2ff7-f1714f9713f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.948163] env[61728]: DEBUG nova.network.neutron [-] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.953011] env[61728]: DEBUG nova.compute.manager [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Received event network-vif-plugged-56270ade-0651-4e82-8fdd-4bcc30bfd53c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1046.953331] env[61728]: DEBUG oslo_concurrency.lockutils [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] Acquiring lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.953552] env[61728]: DEBUG oslo_concurrency.lockutils [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] Lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.953727] env[61728]: DEBUG oslo_concurrency.lockutils [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] Lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.953898] env[61728]: DEBUG nova.compute.manager [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] No waiting events found dispatching network-vif-plugged-56270ade-0651-4e82-8fdd-4bcc30bfd53c {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1046.954113] env[61728]: WARNING nova.compute.manager [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Received unexpected event network-vif-plugged-56270ade-0651-4e82-8fdd-4bcc30bfd53c for instance with vm_state building and task_state spawning. [ 1046.954243] env[61728]: DEBUG nova.compute.manager [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Received event network-changed-56270ade-0651-4e82-8fdd-4bcc30bfd53c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1046.954393] env[61728]: DEBUG nova.compute.manager [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Refreshing instance network info cache due to event network-changed-56270ade-0651-4e82-8fdd-4bcc30bfd53c. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1046.954565] env[61728]: DEBUG oslo_concurrency.lockutils [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] Acquiring lock "refresh_cache-e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.019869] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Releasing lock "refresh_cache-e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.020103] env[61728]: DEBUG nova.compute.manager [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Instance network_info: |[{"id": "56270ade-0651-4e82-8fdd-4bcc30bfd53c", "address": "fa:16:3e:35:57:51", "network": {"id": "11128295-2780-4c0a-81cf-946216b8d623", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-492544146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d88c08b899c94efb8254e510bdf53d25", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56270ade-06", "ovs_interfaceid": "56270ade-0651-4e82-8fdd-4bcc30bfd53c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1047.020402] env[61728]: DEBUG oslo_concurrency.lockutils [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] Acquired lock "refresh_cache-e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.020591] env[61728]: DEBUG nova.network.neutron [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Refreshing network info cache for port 56270ade-0651-4e82-8fdd-4bcc30bfd53c {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1047.021964] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:57:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3ff3baee-99ce-4b51-ae98-efc6163aaab3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '56270ade-0651-4e82-8fdd-4bcc30bfd53c', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1047.029373] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Creating folder: Project (d88c08b899c94efb8254e510bdf53d25). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1047.032758] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b8f93aa3-57ef-4c28-989e-a4ed23e4c425 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.044205] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Created folder: Project (d88c08b899c94efb8254e510bdf53d25) in parent group-v121913. [ 1047.044407] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Creating folder: Instances. Parent ref: group-v122196. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1047.044649] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f6ef2a83-cb12-4346-984c-8076db740d59 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.052743] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Created folder: Instances in parent group-v122196. [ 1047.052975] env[61728]: DEBUG oslo.service.loopingcall [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1047.053195] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1047.053424] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5932f7d6-3e4b-4bdc-96e9-dbc3b55af2c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.073841] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1047.073841] env[61728]: value = "task-464780" [ 1047.073841] env[61728]: _type = "Task" [ 1047.073841] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.082565] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464780, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.184334] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "267880ff-0fac-4763-82be-ef8587ee7628" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.313297] env[61728]: DEBUG nova.compute.manager [req-1d835f95-6f60-4168-b147-11c19a73770f req-91f1f76c-8531-4ec9-94c2-07e4fb40c509 service nova] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Received event network-changed-63b2bf95-c628-4c62-bdbd-051192e13a2f {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1047.313297] env[61728]: DEBUG nova.compute.manager [req-1d835f95-6f60-4168-b147-11c19a73770f req-91f1f76c-8531-4ec9-94c2-07e4fb40c509 service nova] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Refreshing instance network info cache due to event network-changed-63b2bf95-c628-4c62-bdbd-051192e13a2f. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1047.313435] env[61728]: DEBUG oslo_concurrency.lockutils [req-1d835f95-6f60-4168-b147-11c19a73770f req-91f1f76c-8531-4ec9-94c2-07e4fb40c509 service nova] Acquiring lock "refresh_cache-41329d0d-3838-4d1b-a4e6-8df4508eea76" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.313517] env[61728]: DEBUG oslo_concurrency.lockutils [req-1d835f95-6f60-4168-b147-11c19a73770f req-91f1f76c-8531-4ec9-94c2-07e4fb40c509 service nova] Acquired lock "refresh_cache-41329d0d-3838-4d1b-a4e6-8df4508eea76" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.313643] env[61728]: DEBUG nova.network.neutron [req-1d835f95-6f60-4168-b147-11c19a73770f req-91f1f76c-8531-4ec9-94c2-07e4fb40c509 service nova] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Refreshing network info cache for port 63b2bf95-c628-4c62-bdbd-051192e13a2f {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1047.356582] env[61728]: DEBUG nova.network.neutron [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Updated VIF entry in instance network info cache for port 56270ade-0651-4e82-8fdd-4bcc30bfd53c. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1047.356582] env[61728]: DEBUG nova.network.neutron [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Updating instance_info_cache with network_info: [{"id": "56270ade-0651-4e82-8fdd-4bcc30bfd53c", "address": "fa:16:3e:35:57:51", "network": {"id": "11128295-2780-4c0a-81cf-946216b8d623", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-492544146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d88c08b899c94efb8254e510bdf53d25", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56270ade-06", "ovs_interfaceid": "56270ade-0651-4e82-8fdd-4bcc30bfd53c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.399027] env[61728]: DEBUG nova.scheduler.client.report [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1047.412170] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520cd4e5-3e84-52ab-2ff7-f1714f9713f7, 'name': SearchDatastore_Task, 'duration_secs': 0.009474} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.413383] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abe9400f-4901-4ce0-9fea-eb73819891da {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.418946] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1047.418946] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ad8f4a-90cf-7fcd-21db-eca51e8ec258" [ 1047.418946] env[61728]: _type = "Task" [ 1047.418946] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.429016] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ad8f4a-90cf-7fcd-21db-eca51e8ec258, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.452580] env[61728]: INFO nova.compute.manager [-] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Took 1.24 seconds to deallocate network for instance. [ 1047.585148] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464780, 'name': CreateVM_Task, 'duration_secs': 0.329403} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.585148] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1047.586164] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.586518] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.589052] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1047.589052] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de0622b3-f432-4a6a-8954-0c3848a61270 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.594598] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Waiting for the task: (returnval){ [ 1047.594598] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c211d6-3435-4e05-3dfc-5003f928a9b1" [ 1047.594598] env[61728]: _type = "Task" [ 1047.594598] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.602037] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c211d6-3435-4e05-3dfc-5003f928a9b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.860406] env[61728]: DEBUG oslo_concurrency.lockutils [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] Releasing lock "refresh_cache-e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.860406] env[61728]: DEBUG nova.compute.manager [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Received event network-vif-deleted-c8407eac-4337-4ae3-a071-08f48bf87248 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1047.860406] env[61728]: INFO nova.compute.manager [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Neutron deleted interface c8407eac-4337-4ae3-a071-08f48bf87248; detaching it from the instance and deleting it from the info cache [ 1047.860406] env[61728]: DEBUG nova.network.neutron [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.905669] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.810s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.908656] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 5.324s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.932571] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ad8f4a-90cf-7fcd-21db-eca51e8ec258, 'name': SearchDatastore_Task, 'duration_secs': 0.009544} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.932901] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.933375] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 77b8f509-17e7-4c25-a43d-1655c16ce8d1/77b8f509-17e7-4c25-a43d-1655c16ce8d1.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1047.934231] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-80a862d1-93a2-4ad8-8304-c6194eaba589 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.940691] env[61728]: INFO nova.scheduler.client.report [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleted allocations for instance b4241d6f-1d70-4f61-a322-caff7c4e6f64 [ 1047.943498] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1047.943498] env[61728]: value = "task-464781" [ 1047.943498] env[61728]: _type = "Task" [ 1047.943498] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.956951] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464781, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.957660] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.111478] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c211d6-3435-4e05-3dfc-5003f928a9b1, 'name': SearchDatastore_Task, 'duration_secs': 0.009872} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.111527] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.112073] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1048.112151] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.112269] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.112463] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1048.112751] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-52012fe5-ac64-48ed-b624-a0d914586f5c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.126516] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1048.126516] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1048.127604] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f599e7da-8437-4c6f-9b3b-8b9caa895921 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.134933] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Waiting for the task: (returnval){ [ 1048.134933] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529c4aae-119b-e9a6-d043-0676a22083d2" [ 1048.134933] env[61728]: _type = "Task" [ 1048.134933] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.143105] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529c4aae-119b-e9a6-d043-0676a22083d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.273254] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "267880ff-0fac-4763-82be-ef8587ee7628" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.273431] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "267880ff-0fac-4763-82be-ef8587ee7628" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.273668] env[61728]: INFO nova.compute.manager [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Attaching volume 9132f266-e949-4c16-9e9b-57be667f7110 to /dev/sdb [ 1048.330204] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-007649ae-d21d-4054-9f0f-d1c2f04f8651 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.343248] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-665c997c-7bda-405b-b886-a149a3483301 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.371855] env[61728]: DEBUG nova.virt.block_device [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Updating existing volume attachment record: 11eb3106-0a02-4215-a87e-092d2f010069 {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1048.374766] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c4c1c431-e36d-4f04-98e7-a884642a020d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.387222] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854c41aa-25a1-4ae5-8757-8ef46b1204d0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.431880] env[61728]: DEBUG nova.compute.manager [req-263875d8-83b7-4634-aed0-4b35437ff334 req-d202bf65-7dd0-4be0-a575-dc8f1ee48cf6 service nova] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Detach interface failed, port_id=c8407eac-4337-4ae3-a071-08f48bf87248, reason: Instance a70d1c37-e792-4168-b7e1-9418b8cb7818 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1048.432908] env[61728]: DEBUG nova.network.neutron [req-1d835f95-6f60-4168-b147-11c19a73770f req-91f1f76c-8531-4ec9-94c2-07e4fb40c509 service nova] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Updated VIF entry in instance network info cache for port 63b2bf95-c628-4c62-bdbd-051192e13a2f. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1048.433326] env[61728]: DEBUG nova.network.neutron [req-1d835f95-6f60-4168-b147-11c19a73770f req-91f1f76c-8531-4ec9-94c2-07e4fb40c509 service nova] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Updating instance_info_cache with network_info: [{"id": "63b2bf95-c628-4c62-bdbd-051192e13a2f", "address": "fa:16:3e:fc:4c:88", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63b2bf95-c6", "ovs_interfaceid": "63b2bf95-c628-4c62-bdbd-051192e13a2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.458273] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464781, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486107} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.459329] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1ff0fee7-0bea-4217-9c5f-38954d2574a8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "b4241d6f-1d70-4f61-a322-caff7c4e6f64" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.937s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.460421] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 77b8f509-17e7-4c25-a43d-1655c16ce8d1/77b8f509-17e7-4c25-a43d-1655c16ce8d1.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1048.460691] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1048.461171] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c036f1b0-4b68-47b9-a83e-4858ddae2f39 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.468617] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1048.468617] env[61728]: value = "task-464782" [ 1048.468617] env[61728]: _type = "Task" [ 1048.468617] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.477713] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464782, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.652654] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529c4aae-119b-e9a6-d043-0676a22083d2, 'name': SearchDatastore_Task, 'duration_secs': 0.067291} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.653662] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77f1c1ca-3b9e-43e3-944c-aede4f19289a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.660032] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Waiting for the task: (returnval){ [ 1048.660032] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5215a12e-4837-b00e-1b93-53ebf062c4ff" [ 1048.660032] env[61728]: _type = "Task" [ 1048.660032] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.667255] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5215a12e-4837-b00e-1b93-53ebf062c4ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.937909] env[61728]: INFO nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating resource usage from migration 3f7c1af7-63cb-44c6-b2d2-a0146aa543d0 [ 1048.940295] env[61728]: DEBUG oslo_concurrency.lockutils [req-1d835f95-6f60-4168-b147-11c19a73770f req-91f1f76c-8531-4ec9-94c2-07e4fb40c509 service nova] Releasing lock "refresh_cache-41329d0d-3838-4d1b-a4e6-8df4508eea76" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.957754] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 40ef29df-12d6-4274-846a-4a2c32752b8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.957910] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance f27f22aa-f05d-4981-a389-311731c51f93 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.958088] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 5fa2e9f4-6d68-46c2-a549-95d216bab886 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.958224] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance b1fb2e3f-234a-44c7-983e-d4441b4e3cef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.958401] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 8c5b2380-bd14-451e-9612-2767c2a6adb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.958491] env[61728]: WARNING nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance a70d1c37-e792-4168-b7e1-9418b8cb7818 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1048.958610] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.958726] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 267880ff-0fac-4763-82be-ef8587ee7628 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.958842] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance d55beb02-00fc-4df6-a239-9e5d776a7c18 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.958954] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 41329d0d-3838-4d1b-a4e6-8df4508eea76 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.959078] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 77b8f509-17e7-4c25-a43d-1655c16ce8d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.959193] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.959307] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Migration 3f7c1af7-63cb-44c6-b2d2-a0146aa543d0 is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1048.959441] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 6c154517-68e8-4e70-86eb-cf53f6f812d8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.959649] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1048.959790] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2816MB phys_disk=200GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1048.978918] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464782, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.169474} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.981195] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1048.982469] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a3e4f5-5201-493c-aa2a-0d41d0c38729 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.006250] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 77b8f509-17e7-4c25-a43d-1655c16ce8d1/77b8f509-17e7-4c25-a43d-1655c16ce8d1.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.009249] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e87255a-4a7b-4959-b65d-a837f09c6772 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.029970] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1049.029970] env[61728]: value = "task-464784" [ 1049.029970] env[61728]: _type = "Task" [ 1049.029970] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.042031] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464784, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.172312] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5215a12e-4837-b00e-1b93-53ebf062c4ff, 'name': SearchDatastore_Task, 'duration_secs': 0.027521} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.175112] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.175404] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0/e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1049.175604] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e63b4e7-f26c-4a0b-9658-7313c488a373 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.182134] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Waiting for the task: (returnval){ [ 1049.182134] env[61728]: value = "task-464785" [ 1049.182134] env[61728]: _type = "Task" [ 1049.182134] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.186727] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f67986-8519-4ffd-8641-459731e90462 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.195971] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e940b3-1eee-4584-b8b3-9defde4f34a5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.199182] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464785, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.226419] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ef5c72-1b41-4b13-8188-aed3f7619da3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.233680] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e111a33-d2de-42d5-82fd-3b4c66bc22c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.246574] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.544367] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464784, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.694606] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464785, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.749832] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.995277] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.995597] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.041171] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464784, 'name': ReconfigVM_Task, 'duration_secs': 0.803884} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.041458] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 77b8f509-17e7-4c25-a43d-1655c16ce8d1/77b8f509-17e7-4c25-a43d-1655c16ce8d1.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.042112] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2ca4f964-3f57-405d-9fff-78ab80163273 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.048987] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1050.048987] env[61728]: value = "task-464786" [ 1050.048987] env[61728]: _type = "Task" [ 1050.048987] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.057707] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464786, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.193188] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464785, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.640818} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.193523] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0/e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1050.193744] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1050.194054] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ec30a1b7-2485-420b-82cb-fe5eaae694d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.200937] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Waiting for the task: (returnval){ [ 1050.200937] env[61728]: value = "task-464787" [ 1050.200937] env[61728]: _type = "Task" [ 1050.200937] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.208838] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464787, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.255186] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1050.255407] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.348s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.256051] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.793s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.257481] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1050.497528] env[61728]: DEBUG nova.compute.manager [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1050.558998] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464786, 'name': Rename_Task, 'duration_secs': 0.13551} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.559336] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1050.559632] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5049fcac-e1a7-470d-b92d-a6dd842a29aa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.566619] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1050.566619] env[61728]: value = "task-464788" [ 1050.566619] env[61728]: _type = "Task" [ 1050.566619] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.574503] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464788, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.713025] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464787, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061864} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.713025] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1050.713025] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13f5d52-0a81-4786-920c-bd597e9e3ef6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.734503] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0/e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.734907] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-480bad64-267e-443b-80ed-84ac8b1f1818 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.754640] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Waiting for the task: (returnval){ [ 1050.754640] env[61728]: value = "task-464790" [ 1050.754640] env[61728]: _type = "Task" [ 1050.754640] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.760688] env[61728]: INFO nova.compute.claims [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1050.767615] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464790, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.022358] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.077259] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464788, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.266115] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464790, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.270210] env[61728]: INFO nova.compute.resource_tracker [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating resource usage from migration 3f7c1af7-63cb-44c6-b2d2-a0146aa543d0 [ 1051.364919] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.365167] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.365314] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Starting heal instance info cache {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1051.440628] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d80a944-5d34-4a0e-a833-cbbc17d68603 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.448055] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61812dec-43f3-44ef-b789-c53817e5073f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.477324] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9cab63e-5b5f-4807-8f14-d262cce12ccd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.484275] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130e9098-2cc3-48d6-81da-f7c6fbe7dbe3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.496773] env[61728]: DEBUG nova.compute.provider_tree [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.577072] env[61728]: DEBUG oslo_vmware.api [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464788, 'name': PowerOnVM_Task, 'duration_secs': 0.550971} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.577363] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1051.577572] env[61728]: INFO nova.compute.manager [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Took 9.40 seconds to spawn the instance on the hypervisor. [ 1051.577757] env[61728]: DEBUG nova.compute.manager [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1051.578533] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4f4eb8-283f-49b2-a2ec-47208d8a5416 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.764087] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464790, 'name': ReconfigVM_Task, 'duration_secs': 0.55157} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.764382] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Reconfigured VM instance instance-0000006d to attach disk [datastore1] e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0/e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.764991] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-da85c4be-d009-4651-92ea-e6beb7a4af7b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.771073] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Waiting for the task: (returnval){ [ 1051.771073] env[61728]: value = "task-464791" [ 1051.771073] env[61728]: _type = "Task" [ 1051.771073] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.778753] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464791, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.905868] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.906051] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquired lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.906208] env[61728]: DEBUG nova.network.neutron [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Forcefully refreshing network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1051.999595] env[61728]: DEBUG nova.scheduler.client.report [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1052.094221] env[61728]: INFO nova.compute.manager [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Took 17.71 seconds to build instance. [ 1052.281617] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464791, 'name': Rename_Task, 'duration_secs': 0.133303} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.282634] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1052.282634] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22866859-cad5-4106-b6d5-43b2bd85cd02 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.288942] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Waiting for the task: (returnval){ [ 1052.288942] env[61728]: value = "task-464792" [ 1052.288942] env[61728]: _type = "Task" [ 1052.288942] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.296423] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464792, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.393752] env[61728]: DEBUG nova.compute.manager [req-efecd814-3fba-4312-b880-0f80d176861f req-5a039c78-f8d5-4cee-918f-ef374bb454ab service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Received event network-changed-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1052.394014] env[61728]: DEBUG nova.compute.manager [req-efecd814-3fba-4312-b880-0f80d176861f req-5a039c78-f8d5-4cee-918f-ef374bb454ab service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Refreshing instance network info cache due to event network-changed-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1052.394272] env[61728]: DEBUG oslo_concurrency.lockutils [req-efecd814-3fba-4312-b880-0f80d176861f req-5a039c78-f8d5-4cee-918f-ef374bb454ab service nova] Acquiring lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.394422] env[61728]: DEBUG oslo_concurrency.lockutils [req-efecd814-3fba-4312-b880-0f80d176861f req-5a039c78-f8d5-4cee-918f-ef374bb454ab service nova] Acquired lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.394585] env[61728]: DEBUG nova.network.neutron [req-efecd814-3fba-4312-b880-0f80d176861f req-5a039c78-f8d5-4cee-918f-ef374bb454ab service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Refreshing network info cache for port ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1052.504664] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.248s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.504664] env[61728]: INFO nova.compute.manager [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Migrating [ 1052.511660] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.554s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.512419] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.514681] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.492s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.516250] env[61728]: INFO nova.compute.claims [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.538596] env[61728]: INFO nova.scheduler.client.report [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Deleted allocations for instance a70d1c37-e792-4168-b7e1-9418b8cb7818 [ 1052.596207] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5faa8980-77db-4acf-aff3-ca6a7b86c0c4 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.222s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.798925] env[61728]: DEBUG oslo_vmware.api [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464792, 'name': PowerOnVM_Task, 'duration_secs': 0.478221} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.799217] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1052.799423] env[61728]: INFO nova.compute.manager [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Took 8.29 seconds to spawn the instance on the hypervisor. [ 1052.799628] env[61728]: DEBUG nova.compute.manager [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.800392] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a958c6-b209-4228-b19a-6d6355cb4134 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.939462] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1052.939927] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122199', 'volume_id': '9132f266-e949-4c16-9e9b-57be667f7110', 'name': 'volume-9132f266-e949-4c16-9e9b-57be667f7110', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '267880ff-0fac-4763-82be-ef8587ee7628', 'attached_at': '', 'detached_at': '', 'volume_id': '9132f266-e949-4c16-9e9b-57be667f7110', 'serial': '9132f266-e949-4c16-9e9b-57be667f7110'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1052.941176] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ce0597a-3a47-4829-9ffb-eb31b57eeea7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.960301] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762a4f21-ea18-432f-beee-addb4a701c5c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.986034] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] volume-9132f266-e949-4c16-9e9b-57be667f7110/volume-9132f266-e949-4c16-9e9b-57be667f7110.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1052.986390] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b424e0ce-9c74-45d5-b7dc-6b8b677e3fae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.008260] env[61728]: DEBUG oslo_vmware.api [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1053.008260] env[61728]: value = "task-464793" [ 1053.008260] env[61728]: _type = "Task" [ 1053.008260] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.014972] env[61728]: DEBUG oslo_vmware.api [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464793, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.027523] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.027827] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.028158] env[61728]: DEBUG nova.network.neutron [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1053.048394] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc4601c3-9bae-4c32-bafa-147546108fe6 tempest-AttachInterfacesTestJSON-84892408 tempest-AttachInterfacesTestJSON-84892408-project-member] Lock "a70d1c37-e792-4168-b7e1-9418b8cb7818" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.966s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.147805] env[61728]: DEBUG nova.network.neutron [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Updating instance_info_cache with network_info: [{"id": "4a6313df-8741-4533-ae0e-1f469193c389", "address": "fa:16:3e:43:88:0c", "network": {"id": "444d470c-2ff4-42d9-9aa4-35296337df1e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266708519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c035f75f40864559b47b643fcfca56e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a6313df-87", "ovs_interfaceid": "4a6313df-8741-4533-ae0e-1f469193c389", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.158539] env[61728]: DEBUG nova.network.neutron [req-efecd814-3fba-4312-b880-0f80d176861f req-5a039c78-f8d5-4cee-918f-ef374bb454ab service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Updated VIF entry in instance network info cache for port ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1053.158900] env[61728]: DEBUG nova.network.neutron [req-efecd814-3fba-4312-b880-0f80d176861f req-5a039c78-f8d5-4cee-918f-ef374bb454ab service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Updating instance_info_cache with network_info: [{"id": "ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb", "address": "fa:16:3e:36:bc:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea6ae87d-d3", "ovs_interfaceid": "ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.316410] env[61728]: INFO nova.compute.manager [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Took 17.43 seconds to build instance. [ 1053.518056] env[61728]: DEBUG oslo_vmware.api [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464793, 'name': ReconfigVM_Task, 'duration_secs': 0.433579} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.518447] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Reconfigured VM instance instance-00000064 to attach disk [datastore1] volume-9132f266-e949-4c16-9e9b-57be667f7110/volume-9132f266-e949-4c16-9e9b-57be667f7110.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.523766] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2e0420a-e014-4039-8294-eca797af1777 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.544521] env[61728]: DEBUG oslo_vmware.api [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1053.544521] env[61728]: value = "task-464794" [ 1053.544521] env[61728]: _type = "Task" [ 1053.544521] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.554754] env[61728]: DEBUG oslo_vmware.api [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464794, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.650286] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Releasing lock "refresh_cache-f27f22aa-f05d-4981-a389-311731c51f93" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.650435] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Updated the network info_cache for instance {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1053.650682] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.650892] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.651076] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.651311] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.651438] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.651613] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.651802] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61728) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1053.651970] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.661164] env[61728]: DEBUG oslo_concurrency.lockutils [req-efecd814-3fba-4312-b880-0f80d176861f req-5a039c78-f8d5-4cee-918f-ef374bb454ab service nova] Releasing lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.807645] env[61728]: DEBUG nova.network.neutron [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating instance_info_cache with network_info: [{"id": "e7f1c36b-89f0-4858-96a0-3c1875c4bb29", "address": "fa:16:3e:c0:51:65", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7f1c36b-89", "ovs_interfaceid": "e7f1c36b-89f0-4858-96a0-3c1875c4bb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.818609] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b6deca59-4519-4be1-b62a-2c1db46b8f32 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.946s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.891630] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a9acdd-57ad-408f-9a56-69913c12f899 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.900148] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ade1042-b984-432c-92d5-bf0f09a58cd9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.930396] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc59b4e7-2ab3-48ed-b3f9-c273d6d411bb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.937860] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b598e8fa-a4ff-4819-87ae-ede4a9ae53e7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.951631] env[61728]: DEBUG nova.compute.provider_tree [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.055143] env[61728]: DEBUG oslo_vmware.api [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464794, 'name': ReconfigVM_Task, 'duration_secs': 0.263938} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.055498] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122199', 'volume_id': '9132f266-e949-4c16-9e9b-57be667f7110', 'name': 'volume-9132f266-e949-4c16-9e9b-57be667f7110', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '267880ff-0fac-4763-82be-ef8587ee7628', 'attached_at': '', 'detached_at': '', 'volume_id': '9132f266-e949-4c16-9e9b-57be667f7110', 'serial': '9132f266-e949-4c16-9e9b-57be667f7110'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1054.155235] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.293329] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquiring lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.293329] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.293329] env[61728]: INFO nova.compute.manager [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Rebooting instance [ 1054.314341] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.455276] env[61728]: DEBUG nova.scheduler.client.report [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.810156] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquiring lock "refresh_cache-e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.810698] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquired lock "refresh_cache-e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.811174] env[61728]: DEBUG nova.network.neutron [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1054.962018] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.962018] env[61728]: DEBUG nova.compute.manager [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1054.963300] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.808s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.963619] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.963908] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61728) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1054.965111] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c2af87-0de1-4d1d-8aa1-767dfbae00b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.975994] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01e7db4-10e3-4c3f-8ae4-5026ab0610d1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.990867] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ab8bce-527f-42b1-a8b3-04b41e764bce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.998854] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca31adf-a4d3-4d5d-a00b-adccd99ece1e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.033557] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180230MB free_disk=114GB free_vcpus=48 pci_devices=None {{(pid=61728) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1055.033557] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.033796] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.102769] env[61728]: DEBUG nova.objects.instance [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lazy-loading 'flavor' on Instance uuid 267880ff-0fac-4763-82be-ef8587ee7628 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.469256] env[61728]: DEBUG nova.compute.utils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1055.470963] env[61728]: DEBUG nova.compute.manager [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1055.471164] env[61728]: DEBUG nova.network.neutron [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1055.546063] env[61728]: DEBUG nova.network.neutron [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Updating instance_info_cache with network_info: [{"id": "56270ade-0651-4e82-8fdd-4bcc30bfd53c", "address": "fa:16:3e:35:57:51", "network": {"id": "11128295-2780-4c0a-81cf-946216b8d623", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-492544146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d88c08b899c94efb8254e510bdf53d25", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56270ade-06", "ovs_interfaceid": "56270ade-0651-4e82-8fdd-4bcc30bfd53c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.548144] env[61728]: DEBUG nova.policy [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54c30aeeea3a4c1b8ed34430361c2a4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a02bd852c564acd99e12fd17279f028', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1055.608675] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b4f04594-4879-4cc3-b4eb-5ef7ed90b42c tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "267880ff-0fac-4763-82be-ef8587ee7628" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.335s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.834601] env[61728]: DEBUG nova.network.neutron [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Successfully created port: 96bfe2c5-12a3-4a70-b2d9-aab955fb6b08 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1055.837121] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "267880ff-0fac-4763-82be-ef8587ee7628" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.837762] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "267880ff-0fac-4763-82be-ef8587ee7628" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.846223] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284017cf-3ae5-42b7-ade8-242565b0555d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.868278] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating instance '6c154517-68e8-4e70-86eb-cf53f6f812d8' progress to 0 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1055.975712] env[61728]: DEBUG nova.compute.manager [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1056.051350] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Releasing lock "refresh_cache-e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.053312] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Applying migration context for instance 6c154517-68e8-4e70-86eb-cf53f6f812d8 as it has an incoming, in-progress migration 3f7c1af7-63cb-44c6-b2d2-a0146aa543d0. Migration status is migrating {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1056.055333] env[61728]: INFO nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating resource usage from migration 3f7c1af7-63cb-44c6-b2d2-a0146aa543d0 [ 1056.058190] env[61728]: DEBUG nova.compute.manager [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1056.059038] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02643a4-25fe-4886-91c5-76259165934b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.077802] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 40ef29df-12d6-4274-846a-4a2c32752b8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.077990] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance f27f22aa-f05d-4981-a389-311731c51f93 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.078131] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 5fa2e9f4-6d68-46c2-a549-95d216bab886 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.078581] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance b1fb2e3f-234a-44c7-983e-d4441b4e3cef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.078581] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 8c5b2380-bd14-451e-9612-2767c2a6adb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.078581] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.078696] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 267880ff-0fac-4763-82be-ef8587ee7628 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.078806] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance d55beb02-00fc-4df6-a239-9e5d776a7c18 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.078921] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 41329d0d-3838-4d1b-a4e6-8df4508eea76 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.079113] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 77b8f509-17e7-4c25-a43d-1655c16ce8d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.079213] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.079367] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Migration 3f7c1af7-63cb-44c6-b2d2-a0146aa543d0 is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1056.079427] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 6c154517-68e8-4e70-86eb-cf53f6f812d8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.079827] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.079827] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1056.079827] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3264MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1056.278713] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4877c590-a634-4c4c-b017-6c81c7bb3ccb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.289040] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691e7312-8dc3-41aa-8dfd-e8213f2a1555 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.325311] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732702de-e42c-468c-8df2-dfac877543ed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.333176] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ca9ecb-dc46-40ab-b09c-e015d1bd9a4e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.346983] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.350332] env[61728]: INFO nova.compute.manager [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Detaching volume 9132f266-e949-4c16-9e9b-57be667f7110 [ 1056.376315] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1056.376580] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b75fa91-6799-482b-8f94-495d7cd8af5c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.383554] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1056.383554] env[61728]: value = "task-464795" [ 1056.383554] env[61728]: _type = "Task" [ 1056.383554] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.392156] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464795, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.392855] env[61728]: INFO nova.virt.block_device [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Attempting to driver detach volume 9132f266-e949-4c16-9e9b-57be667f7110 from mountpoint /dev/sdb [ 1056.393091] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1056.393290] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122199', 'volume_id': '9132f266-e949-4c16-9e9b-57be667f7110', 'name': 'volume-9132f266-e949-4c16-9e9b-57be667f7110', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '267880ff-0fac-4763-82be-ef8587ee7628', 'attached_at': '', 'detached_at': '', 'volume_id': '9132f266-e949-4c16-9e9b-57be667f7110', 'serial': '9132f266-e949-4c16-9e9b-57be667f7110'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1056.394061] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937c9cd5-8a2b-411e-940e-65190b9258fb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.414710] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074f5557-5960-43d5-b528-016dc2fab3d2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.421340] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35eb930d-442b-4288-b753-d542058e8955 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.442985] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9abe3626-703c-45b5-bf6e-1e9de6e5b6e8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.458196] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] The volume has not been displaced from its original location: [datastore1] volume-9132f266-e949-4c16-9e9b-57be667f7110/volume-9132f266-e949-4c16-9e9b-57be667f7110.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1056.463690] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Reconfiguring VM instance instance-00000064 to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1056.463995] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e3f1d22-8242-4213-a1b6-0b7bc7ef65c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.485075] env[61728]: DEBUG oslo_vmware.api [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1056.485075] env[61728]: value = "task-464796" [ 1056.485075] env[61728]: _type = "Task" [ 1056.485075] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.493315] env[61728]: DEBUG oslo_vmware.api [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464796, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.850491] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1056.895568] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464795, 'name': PowerOffVM_Task, 'duration_secs': 0.273858} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.895925] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1056.896185] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating instance '6c154517-68e8-4e70-86eb-cf53f6f812d8' progress to 17 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1056.984314] env[61728]: DEBUG nova.compute.manager [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1056.996346] env[61728]: DEBUG oslo_vmware.api [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464796, 'name': ReconfigVM_Task, 'duration_secs': 0.230307} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.996540] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Reconfigured VM instance instance-00000064 to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1057.001696] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8cdd3525-ec13-40f5-82b1-754854cc73ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.018880] env[61728]: DEBUG oslo_vmware.api [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1057.018880] env[61728]: value = "task-464797" [ 1057.018880] env[61728]: _type = "Task" [ 1057.018880] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.021046] env[61728]: DEBUG nova.virt.hardware [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1057.021289] env[61728]: DEBUG nova.virt.hardware [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1057.021452] env[61728]: DEBUG nova.virt.hardware [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1057.021644] env[61728]: DEBUG nova.virt.hardware [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1057.021841] env[61728]: DEBUG nova.virt.hardware [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1057.022053] env[61728]: DEBUG nova.virt.hardware [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1057.022236] env[61728]: DEBUG nova.virt.hardware [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1057.022406] env[61728]: DEBUG nova.virt.hardware [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1057.022564] env[61728]: DEBUG nova.virt.hardware [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1057.022758] env[61728]: DEBUG nova.virt.hardware [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1057.022989] env[61728]: DEBUG nova.virt.hardware [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1057.023849] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce52cc1a-e413-4066-b812-01b54e74727e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.036763] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ed6f201-1465-4825-9a4c-5287d0a7158a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.040668] env[61728]: DEBUG oslo_vmware.api [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464797, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.076253] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af506c9-beac-4f26-952f-90d388ec03ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.082719] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Doing hard reboot of VM {{(pid=61728) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 1057.082981] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-308020f1-36a9-4ce2-b979-65f4ba8a6972 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.090130] env[61728]: DEBUG oslo_vmware.api [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Waiting for the task: (returnval){ [ 1057.090130] env[61728]: value = "task-464798" [ 1057.090130] env[61728]: _type = "Task" [ 1057.090130] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.098451] env[61728]: DEBUG oslo_vmware.api [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464798, 'name': ResetVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.272942] env[61728]: DEBUG nova.compute.manager [req-38209c56-4142-49b9-881f-6f42c39c747d req-1078d244-16ac-4922-b2c8-907215705418 service nova] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Received event network-vif-plugged-96bfe2c5-12a3-4a70-b2d9-aab955fb6b08 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1057.273432] env[61728]: DEBUG oslo_concurrency.lockutils [req-38209c56-4142-49b9-881f-6f42c39c747d req-1078d244-16ac-4922-b2c8-907215705418 service nova] Acquiring lock "3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.273794] env[61728]: DEBUG oslo_concurrency.lockutils [req-38209c56-4142-49b9-881f-6f42c39c747d req-1078d244-16ac-4922-b2c8-907215705418 service nova] Lock "3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.274835] env[61728]: DEBUG oslo_concurrency.lockutils [req-38209c56-4142-49b9-881f-6f42c39c747d req-1078d244-16ac-4922-b2c8-907215705418 service nova] Lock "3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.275227] env[61728]: DEBUG nova.compute.manager [req-38209c56-4142-49b9-881f-6f42c39c747d req-1078d244-16ac-4922-b2c8-907215705418 service nova] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] No waiting events found dispatching network-vif-plugged-96bfe2c5-12a3-4a70-b2d9-aab955fb6b08 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1057.275563] env[61728]: WARNING nova.compute.manager [req-38209c56-4142-49b9-881f-6f42c39c747d req-1078d244-16ac-4922-b2c8-907215705418 service nova] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Received unexpected event network-vif-plugged-96bfe2c5-12a3-4a70-b2d9-aab955fb6b08 for instance with vm_state building and task_state spawning. [ 1057.356703] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1057.356914] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.323s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.357167] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.357318] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Cleaning up deleted instances {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1057.369281] env[61728]: DEBUG nova.network.neutron [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Successfully updated port: 96bfe2c5-12a3-4a70-b2d9-aab955fb6b08 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1057.402686] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1057.402966] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1057.403157] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1057.403350] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1057.403506] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1057.403663] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1057.404563] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1057.404823] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1057.405060] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1057.405261] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1057.405440] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1057.411206] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dce81590-0070-4a8c-bca7-51419efc3ceb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.429246] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1057.429246] env[61728]: value = "task-464799" [ 1057.429246] env[61728]: _type = "Task" [ 1057.429246] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.435977] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464799, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.535019] env[61728]: DEBUG oslo_vmware.api [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464797, 'name': ReconfigVM_Task, 'duration_secs': 0.175132} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.535019] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122199', 'volume_id': '9132f266-e949-4c16-9e9b-57be667f7110', 'name': 'volume-9132f266-e949-4c16-9e9b-57be667f7110', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '267880ff-0fac-4763-82be-ef8587ee7628', 'attached_at': '', 'detached_at': '', 'volume_id': '9132f266-e949-4c16-9e9b-57be667f7110', 'serial': '9132f266-e949-4c16-9e9b-57be667f7110'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1057.600174] env[61728]: DEBUG oslo_vmware.api [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464798, 'name': ResetVM_Task, 'duration_secs': 0.100623} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.600456] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Did hard reboot of VM {{(pid=61728) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 1057.600644] env[61728]: DEBUG nova.compute.manager [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1057.601428] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f2c86d-cee2-4eac-9f1e-984ec37d6211 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.873913] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] There are 56 instances to clean {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1057.874188] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 5f248cf6-31a9-49ef-a8a8-218987664af7] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1057.876278] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "refresh_cache-3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.876450] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "refresh_cache-3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.876579] env[61728]: DEBUG nova.network.neutron [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1057.937876] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.085039] env[61728]: DEBUG nova.objects.instance [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lazy-loading 'flavor' on Instance uuid 267880ff-0fac-4763-82be-ef8587ee7628 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.116960] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3e6f8d4d-c51e-491e-b33d-70fb4126aad4 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.825s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.379982] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 9efdee89-4c88-4af3-af04-5281e8f6bc25] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1058.438841] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464799, 'name': ReconfigVM_Task, 'duration_secs': 0.881319} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.439365] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating instance '6c154517-68e8-4e70-86eb-cf53f6f812d8' progress to 33 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1058.467978] env[61728]: DEBUG nova.network.neutron [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1058.679440] env[61728]: DEBUG nova.network.neutron [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Updating instance_info_cache with network_info: [{"id": "96bfe2c5-12a3-4a70-b2d9-aab955fb6b08", "address": "fa:16:3e:31:78:cd", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96bfe2c5-12", "ovs_interfaceid": "96bfe2c5-12a3-4a70-b2d9-aab955fb6b08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.888927] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: bb7b38d5-53c7-4c05-96f8-9121171d3c46] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1058.947306] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1058.947618] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1058.947755] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1058.947963] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1058.948321] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1058.948496] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1058.948713] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1058.948879] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1058.949064] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1058.949254] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1058.949436] env[61728]: DEBUG nova.virt.hardware [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1058.955722] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1058.956439] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7cd64b1-b94a-4b0a-b360-f3d035037d27 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.978499] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1058.978499] env[61728]: value = "task-464800" [ 1058.978499] env[61728]: _type = "Task" [ 1058.978499] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.993568] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464800, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.087652] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquiring lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.087953] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.088253] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquiring lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.089253] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.089253] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.091829] env[61728]: INFO nova.compute.manager [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Terminating instance [ 1059.097367] env[61728]: DEBUG nova.compute.manager [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1059.097367] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1059.098077] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2b79d6-c2ba-4a8e-8241-7637ae3f8d4d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.101942] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4921a5e2-3bf1-4176-86a3-1f2ae670f8fa tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "267880ff-0fac-4763-82be-ef8587ee7628" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.264s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.109412] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1059.109412] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8cd38830-6941-42c1-8560-f6a49626f81b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.116207] env[61728]: DEBUG oslo_vmware.api [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Waiting for the task: (returnval){ [ 1059.116207] env[61728]: value = "task-464801" [ 1059.116207] env[61728]: _type = "Task" [ 1059.116207] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.126332] env[61728]: DEBUG oslo_vmware.api [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464801, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.183013] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "refresh_cache-3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.183374] env[61728]: DEBUG nova.compute.manager [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Instance network_info: |[{"id": "96bfe2c5-12a3-4a70-b2d9-aab955fb6b08", "address": "fa:16:3e:31:78:cd", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96bfe2c5-12", "ovs_interfaceid": "96bfe2c5-12a3-4a70-b2d9-aab955fb6b08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1059.183873] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:78:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc6d5964-1106-4345-a26d-185dabd4ff0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '96bfe2c5-12a3-4a70-b2d9-aab955fb6b08', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1059.193763] env[61728]: DEBUG oslo.service.loopingcall [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.194853] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1059.194853] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a23bc8d9-c910-4b86-980d-2c6514bbcf5e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.217032] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1059.217032] env[61728]: value = "task-464802" [ 1059.217032] env[61728]: _type = "Task" [ 1059.217032] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.225277] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464802, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.305247] env[61728]: DEBUG nova.compute.manager [req-ac80b72d-11f3-44a4-9070-f97cb5779c90 req-631b5131-d03b-4f0d-a78b-5ae66203ac98 service nova] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Received event network-changed-96bfe2c5-12a3-4a70-b2d9-aab955fb6b08 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1059.305247] env[61728]: DEBUG nova.compute.manager [req-ac80b72d-11f3-44a4-9070-f97cb5779c90 req-631b5131-d03b-4f0d-a78b-5ae66203ac98 service nova] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Refreshing instance network info cache due to event network-changed-96bfe2c5-12a3-4a70-b2d9-aab955fb6b08. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1059.305449] env[61728]: DEBUG oslo_concurrency.lockutils [req-ac80b72d-11f3-44a4-9070-f97cb5779c90 req-631b5131-d03b-4f0d-a78b-5ae66203ac98 service nova] Acquiring lock "refresh_cache-3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.305595] env[61728]: DEBUG oslo_concurrency.lockutils [req-ac80b72d-11f3-44a4-9070-f97cb5779c90 req-631b5131-d03b-4f0d-a78b-5ae66203ac98 service nova] Acquired lock "refresh_cache-3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.305766] env[61728]: DEBUG nova.network.neutron [req-ac80b72d-11f3-44a4-9070-f97cb5779c90 req-631b5131-d03b-4f0d-a78b-5ae66203ac98 service nova] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Refreshing network info cache for port 96bfe2c5-12a3-4a70-b2d9-aab955fb6b08 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1059.392658] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 378ef752-6d47-4861-baf1-064b2a180dcc] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1059.490329] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464800, 'name': ReconfigVM_Task, 'duration_secs': 0.173141} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.490606] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1059.491473] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-509291be-0461-454b-9bf5-01309e0cc78a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.516145] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 6c154517-68e8-4e70-86eb-cf53f6f812d8/6c154517-68e8-4e70-86eb-cf53f6f812d8.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1059.516441] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03403608-c534-4168-8884-87db6a410fa1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.536207] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1059.536207] env[61728]: value = "task-464803" [ 1059.536207] env[61728]: _type = "Task" [ 1059.536207] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.545916] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464803, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.629338] env[61728]: DEBUG oslo_vmware.api [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464801, 'name': PowerOffVM_Task, 'duration_secs': 0.226552} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.629419] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1059.629565] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1059.630667] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b497e1af-219e-4685-8b15-8e86a7db7500 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.693059] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1059.693423] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1059.693466] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Deleting the datastore file [datastore1] e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.693733] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aaa6425a-10ac-4707-a743-523d0aa4dac2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.700285] env[61728]: DEBUG oslo_vmware.api [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Waiting for the task: (returnval){ [ 1059.700285] env[61728]: value = "task-464805" [ 1059.700285] env[61728]: _type = "Task" [ 1059.700285] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.708496] env[61728]: DEBUG oslo_vmware.api [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464805, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.727280] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464802, 'name': CreateVM_Task, 'duration_secs': 0.369238} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.727466] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1059.728265] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.728418] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.728736] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1059.729102] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a631e9b-e34b-4a1a-be1a-f401d6cb9331 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.733507] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1059.733507] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523ca4e6-f5aa-03fd-2e1a-6eaab6bf1400" [ 1059.733507] env[61728]: _type = "Task" [ 1059.733507] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.741344] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523ca4e6-f5aa-03fd-2e1a-6eaab6bf1400, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.896740] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: b4241d6f-1d70-4f61-a322-caff7c4e6f64] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1060.033811] env[61728]: DEBUG nova.network.neutron [req-ac80b72d-11f3-44a4-9070-f97cb5779c90 req-631b5131-d03b-4f0d-a78b-5ae66203ac98 service nova] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Updated VIF entry in instance network info cache for port 96bfe2c5-12a3-4a70-b2d9-aab955fb6b08. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1060.034351] env[61728]: DEBUG nova.network.neutron [req-ac80b72d-11f3-44a4-9070-f97cb5779c90 req-631b5131-d03b-4f0d-a78b-5ae66203ac98 service nova] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Updating instance_info_cache with network_info: [{"id": "96bfe2c5-12a3-4a70-b2d9-aab955fb6b08", "address": "fa:16:3e:31:78:cd", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96bfe2c5-12", "ovs_interfaceid": "96bfe2c5-12a3-4a70-b2d9-aab955fb6b08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.047071] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464803, 'name': ReconfigVM_Task, 'duration_secs': 0.333018} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.048450] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 6c154517-68e8-4e70-86eb-cf53f6f812d8/6c154517-68e8-4e70-86eb-cf53f6f812d8.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1060.048450] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating instance '6c154517-68e8-4e70-86eb-cf53f6f812d8' progress to 50 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1060.108773] env[61728]: DEBUG oslo_concurrency.lockutils [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "267880ff-0fac-4763-82be-ef8587ee7628" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.109092] env[61728]: DEBUG oslo_concurrency.lockutils [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "267880ff-0fac-4763-82be-ef8587ee7628" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.109304] env[61728]: DEBUG oslo_concurrency.lockutils [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "267880ff-0fac-4763-82be-ef8587ee7628-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.109626] env[61728]: DEBUG oslo_concurrency.lockutils [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "267880ff-0fac-4763-82be-ef8587ee7628-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.109799] env[61728]: DEBUG oslo_concurrency.lockutils [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "267880ff-0fac-4763-82be-ef8587ee7628-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.111809] env[61728]: INFO nova.compute.manager [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Terminating instance [ 1060.113621] env[61728]: DEBUG nova.compute.manager [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1060.113831] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1060.114682] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8c0616-31d2-4d8c-b971-e50cdbaa41b0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.123169] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1060.123722] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f993c00-1958-47dd-926a-4fe41bfd2784 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.130261] env[61728]: DEBUG oslo_vmware.api [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1060.130261] env[61728]: value = "task-464806" [ 1060.130261] env[61728]: _type = "Task" [ 1060.130261] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.138679] env[61728]: DEBUG oslo_vmware.api [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464806, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.212017] env[61728]: DEBUG oslo_vmware.api [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Task: {'id': task-464805, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183006} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.212017] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1060.212017] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1060.212017] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1060.212017] env[61728]: INFO nova.compute.manager [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1060.212017] env[61728]: DEBUG oslo.service.loopingcall [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.212017] env[61728]: DEBUG nova.compute.manager [-] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1060.212017] env[61728]: DEBUG nova.network.neutron [-] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1060.246400] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523ca4e6-f5aa-03fd-2e1a-6eaab6bf1400, 'name': SearchDatastore_Task, 'duration_secs': 0.018512} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.246796] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.247231] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1060.247536] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.247743] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.247984] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1060.248322] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c2fa3a8-3166-4bfa-9dcc-165c176bb2f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.259905] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1060.259905] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1060.260408] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d150cb4a-94b7-4165-9a7e-4d40c2d832cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.269610] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1060.269610] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52841182-cc0f-925f-4aac-63bdcd798262" [ 1060.269610] env[61728]: _type = "Task" [ 1060.269610] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.278270] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52841182-cc0f-925f-4aac-63bdcd798262, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.401791] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 172129ff-136e-4855-8659-3f1ac88dcbd2] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1060.536980] env[61728]: DEBUG oslo_concurrency.lockutils [req-ac80b72d-11f3-44a4-9070-f97cb5779c90 req-631b5131-d03b-4f0d-a78b-5ae66203ac98 service nova] Releasing lock "refresh_cache-3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.556759] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c839bb71-fdee-4104-9fa9-7fcb14c582d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.576500] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622947be-9c9c-4056-aee7-e818ea01562b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.598544] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating instance '6c154517-68e8-4e70-86eb-cf53f6f812d8' progress to 67 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1060.643269] env[61728]: DEBUG oslo_vmware.api [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464806, 'name': PowerOffVM_Task, 'duration_secs': 0.325656} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.643269] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1060.643269] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1060.643269] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c4028f4b-5e88-4ef6-92de-810272b163c3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.707084] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1060.707379] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1060.707508] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Deleting the datastore file [datastore1] 267880ff-0fac-4763-82be-ef8587ee7628 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1060.707785] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-213504f1-1fca-47c4-a934-fe37c3494b7c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.715070] env[61728]: DEBUG oslo_vmware.api [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1060.715070] env[61728]: value = "task-464808" [ 1060.715070] env[61728]: _type = "Task" [ 1060.715070] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.723410] env[61728]: DEBUG oslo_vmware.api [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464808, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.779272] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52841182-cc0f-925f-4aac-63bdcd798262, 'name': SearchDatastore_Task, 'duration_secs': 0.017136} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.780157] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7272384e-de7a-4d8c-8713-c785a2f17b6e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.785865] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1060.785865] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52726013-59bc-0655-1d6c-b6e33dcf7af2" [ 1060.785865] env[61728]: _type = "Task" [ 1060.785865] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.793877] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52726013-59bc-0655-1d6c-b6e33dcf7af2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.908450] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 0edd1f56-3e8e-4868-9c8e-a8caaf026e7f] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1060.982450] env[61728]: DEBUG nova.network.neutron [-] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.152058] env[61728]: DEBUG nova.network.neutron [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Port e7f1c36b-89f0-4858-96a0-3c1875c4bb29 binding to destination host cpu-1 is already ACTIVE {{(pid=61728) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1061.225346] env[61728]: DEBUG oslo_vmware.api [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464808, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.327177} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.225656] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1061.225852] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1061.226083] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1061.226318] env[61728]: INFO nova.compute.manager [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1061.226583] env[61728]: DEBUG oslo.service.loopingcall [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1061.226808] env[61728]: DEBUG nova.compute.manager [-] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1061.226907] env[61728]: DEBUG nova.network.neutron [-] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1061.264156] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "532507fa-996d-45c7-bf70-f09c93be79ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.264411] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "532507fa-996d-45c7-bf70-f09c93be79ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.300267] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52726013-59bc-0655-1d6c-b6e33dcf7af2, 'name': SearchDatastore_Task, 'duration_secs': 0.010269} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.301258] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.301258] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4/3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1061.301258] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2893089-e074-4fed-b59c-6b5119ce60f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.307321] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1061.307321] env[61728]: value = "task-464809" [ 1061.307321] env[61728]: _type = "Task" [ 1061.307321] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.315835] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464809, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.331227] env[61728]: DEBUG nova.compute.manager [req-87dd00e0-3917-4135-984f-6bb8b974ef46 req-96e34b40-9f6f-4143-8cc9-ef26e8a6c5d3 service nova] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Received event network-vif-deleted-56270ade-0651-4e82-8fdd-4bcc30bfd53c {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1061.411610] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 705c8838-f605-46fa-b036-ec212f673f66] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1061.485590] env[61728]: INFO nova.compute.manager [-] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Took 1.27 seconds to deallocate network for instance. [ 1061.767360] env[61728]: DEBUG nova.compute.manager [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1061.818556] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464809, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.915597] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 7e8cde9c-890c-4f50-9ccb-46aa71220395] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1061.994576] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.994841] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.995088] env[61728]: DEBUG nova.objects.instance [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Lazy-loading 'resources' on Instance uuid e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1062.162442] env[61728]: DEBUG nova.network.neutron [-] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.181269] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "6c154517-68e8-4e70-86eb-cf53f6f812d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.181644] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "6c154517-68e8-4e70-86eb-cf53f6f812d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.181952] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "6c154517-68e8-4e70-86eb-cf53f6f812d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.295136] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.318725] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464809, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.419829] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 0ff847f9-97d0-40d9-a458-038fe36b5c62] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1062.668226] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762bc6b4-a8df-4042-ba9d-5a430b4af57e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.672550] env[61728]: INFO nova.compute.manager [-] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Took 1.45 seconds to deallocate network for instance. [ 1062.681028] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c04eac-9e9a-41e3-9dc3-b427fec8b07c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.715373] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0c93a2-e7cb-47bb-9bfa-3c30af4419fe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.723232] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594e2cf6-ae8e-424f-ba7b-3011426f0624 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.737023] env[61728]: DEBUG nova.compute.provider_tree [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.818655] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464809, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.922918] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: db547fea-4d44-440e-ac3e-29ae9b2c547b] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1063.182341] env[61728]: DEBUG oslo_concurrency.lockutils [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.225330] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.225556] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.225748] env[61728]: DEBUG nova.network.neutron [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1063.239983] env[61728]: DEBUG nova.scheduler.client.report [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1063.321297] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464809, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.808213} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.321562] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4/3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1063.322481] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1063.322481] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-49e77468-d8b9-40c6-a5b7-c3b70358c459 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.327625] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1063.327625] env[61728]: value = "task-464810" [ 1063.327625] env[61728]: _type = "Task" [ 1063.327625] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.335027] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464810, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.358773] env[61728]: DEBUG nova.compute.manager [req-37db6212-f630-4af5-ac7a-94e3ca0da656 req-de0b97ec-6df2-44cb-9c10-d48458d414d6 service nova] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Received event network-vif-deleted-eb281159-3780-4ea6-b463-cd8d839be907 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1063.426668] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 0af88a53-bbe7-443e-a7dc-734e483bb730] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1063.745436] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.750s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.747920] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.453s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.750110] env[61728]: INFO nova.compute.claims [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1063.766027] env[61728]: INFO nova.scheduler.client.report [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Deleted allocations for instance e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0 [ 1063.837810] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464810, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06292} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.838122] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1063.838889] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c81643ea-b6cd-48d7-8f78-f84ecb89f57e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.864711] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4/3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.864865] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fdfdca1-e9cc-4639-be45-5f384a9454fa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.888659] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1063.888659] env[61728]: value = "task-464811" [ 1063.888659] env[61728]: _type = "Task" [ 1063.888659] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.897150] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464811, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.930107] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: a70d1c37-e792-4168-b7e1-9418b8cb7818] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1063.951069] env[61728]: DEBUG nova.network.neutron [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating instance_info_cache with network_info: [{"id": "e7f1c36b-89f0-4858-96a0-3c1875c4bb29", "address": "fa:16:3e:c0:51:65", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7f1c36b-89", "ovs_interfaceid": "e7f1c36b-89f0-4858-96a0-3c1875c4bb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.273841] env[61728]: DEBUG oslo_concurrency.lockutils [None req-88443014-9326-4c18-b16a-d3ce66840d77 tempest-InstanceActionsTestJSON-338096335 tempest-InstanceActionsTestJSON-338096335-project-member] Lock "e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.186s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.399378] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464811, 'name': ReconfigVM_Task, 'duration_secs': 0.284384} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.399611] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4/3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.400267] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-11e95cae-7ce3-416d-a649-d6d847e29ef6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.406326] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1064.406326] env[61728]: value = "task-464812" [ 1064.406326] env[61728]: _type = "Task" [ 1064.406326] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.415152] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464812, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.433207] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: aebc4c7c-09d6-484f-8ad6-2416b3a41828] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1064.454451] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.916230] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464812, 'name': Rename_Task, 'duration_secs': 0.16416} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.919346] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1064.919543] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1eaa4fd-65bb-4a0c-8b2a-0c95a5c434c5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.926351] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1064.926351] env[61728]: value = "task-464813" [ 1064.926351] env[61728]: _type = "Task" [ 1064.926351] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.935600] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464813, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.937044] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: ab79ce8c-d2e4-4027-9f0c-87b7fb3dfa69] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1064.979276] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfebf7b3-fc13-425a-a9d8-1daa3d0e48ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.984752] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34d06ac-c9b5-4c6d-990e-3f56cfe49267 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.007984] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b6db02-e806-4bd7-a354-2dc1e1f22d91 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.013496] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c97dd26f-8ba7-4dda-8b31-eb92879b48f8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.022952] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating instance '6c154517-68e8-4e70-86eb-cf53f6f812d8' progress to 83 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1065.053148] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75bb6629-1d76-4d2b-bcc2-59781c5ace15 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.061656] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d53f76a-0e2e-4902-8341-78ec8d417e0a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.076705] env[61728]: DEBUG nova.compute.provider_tree [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.436426] env[61728]: DEBUG oslo_vmware.api [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464813, 'name': PowerOnVM_Task, 'duration_secs': 0.441035} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.436747] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1065.437101] env[61728]: INFO nova.compute.manager [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Took 8.45 seconds to spawn the instance on the hypervisor. [ 1065.437249] env[61728]: DEBUG nova.compute.manager [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.437981] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc75397-4eaf-4b77-aa96-6b657af2688b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.440746] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 2c166dd3-985c-4637-92ab-939b46a7c90e] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1065.555020] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1065.555450] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65b561f0-5912-43f4-a390-ac7da29e4458 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.562848] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1065.562848] env[61728]: value = "task-464814" [ 1065.562848] env[61728]: _type = "Task" [ 1065.562848] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.571276] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464814, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.579283] env[61728]: DEBUG nova.scheduler.client.report [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1065.944793] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 0e77ef23-3d1a-4f00-83f9-5b617334f980] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1065.960040] env[61728]: INFO nova.compute.manager [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Took 14.96 seconds to build instance. [ 1066.073155] env[61728]: DEBUG oslo_vmware.api [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464814, 'name': PowerOnVM_Task, 'duration_secs': 0.410099} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.073431] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1066.073621] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2accba39-423a-43a7-963c-63d9f7e63808 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating instance '6c154517-68e8-4e70-86eb-cf53f6f812d8' progress to 100 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1066.084719] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.337s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.085229] env[61728]: DEBUG nova.compute.manager [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1066.087767] env[61728]: DEBUG oslo_concurrency.lockutils [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.906s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.089766] env[61728]: DEBUG nova.objects.instance [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lazy-loading 'resources' on Instance uuid 267880ff-0fac-4763-82be-ef8587ee7628 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.449330] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 2400abb1-8a07-4f6b-8818-778a2105beb8] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1066.462652] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3ee4678-f0ca-4c5d-bdb6-9c44457df058 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.467s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.590988] env[61728]: DEBUG nova.compute.utils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1066.595251] env[61728]: DEBUG nova.compute.manager [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1066.595437] env[61728]: DEBUG nova.network.neutron [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1066.660785] env[61728]: DEBUG nova.policy [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '836bc29cb90f4b3fb2459041a65805c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '438de94d086e47b99cedc6e4a33471bf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1066.753151] env[61728]: DEBUG oslo_concurrency.lockutils [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.753456] env[61728]: DEBUG oslo_concurrency.lockutils [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.785054] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bef002a-80c8-4405-b5af-a6a562545b53 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.793742] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a46d413b-ffbb-475b-94b4-861f5e3d82d1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.826542] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e229f6f9-da9f-4a23-97cd-8f05ede7ec2b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.833945] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ddd967-8ef5-46d0-84fa-3674722b14ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.847176] env[61728]: DEBUG nova.compute.provider_tree [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1066.951541] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: ae25dc25-f600-471f-aaaa-f166ed691bf3] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1067.101285] env[61728]: DEBUG nova.compute.manager [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1067.111532] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.115028] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.115028] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.115028] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.115028] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.116221] env[61728]: INFO nova.compute.manager [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Terminating instance [ 1067.119497] env[61728]: DEBUG nova.compute.manager [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1067.119664] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1067.120739] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1593574b-98e2-4c07-9295-14405ccc20e9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.140647] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1067.140647] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b0676428-1e6e-46f2-8f15-30ee61390317 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.150020] env[61728]: DEBUG oslo_vmware.api [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1067.150020] env[61728]: value = "task-464815" [ 1067.150020] env[61728]: _type = "Task" [ 1067.150020] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.159913] env[61728]: DEBUG oslo_vmware.api [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464815, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.168967] env[61728]: DEBUG nova.network.neutron [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Successfully created port: d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1067.256745] env[61728]: INFO nova.compute.manager [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Detaching volume 9e2549ce-9f21-4a53-af77-978682bfcbbe [ 1067.312034] env[61728]: INFO nova.virt.block_device [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Attempting to driver detach volume 9e2549ce-9f21-4a53-af77-978682bfcbbe from mountpoint /dev/sdb [ 1067.312034] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1067.312034] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122182', 'volume_id': '9e2549ce-9f21-4a53-af77-978682bfcbbe', 'name': 'volume-9e2549ce-9f21-4a53-af77-978682bfcbbe', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a', 'attached_at': '', 'detached_at': '', 'volume_id': '9e2549ce-9f21-4a53-af77-978682bfcbbe', 'serial': '9e2549ce-9f21-4a53-af77-978682bfcbbe'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1067.312774] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34b8ae4-e838-46a4-8f92-5c314f6ee31d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.336488] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3672e1-5226-448f-b603-6de9453f04f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.344419] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67fc9ffd-be32-46ed-846e-4cfb6f16bd72 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.368058] env[61728]: DEBUG nova.scheduler.client.report [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1067.373030] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9030a43c-5b2c-473f-83ab-d7852c917125 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.392497] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] The volume has not been displaced from its original location: [datastore1] volume-9e2549ce-9f21-4a53-af77-978682bfcbbe/volume-9e2549ce-9f21-4a53-af77-978682bfcbbe.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1067.398329] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Reconfiguring VM instance instance-0000005c to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1067.399297] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5cfb8d7-1e0d-4cc2-93af-8e9b5fb34368 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.419036] env[61728]: DEBUG oslo_vmware.api [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1067.419036] env[61728]: value = "task-464816" [ 1067.419036] env[61728]: _type = "Task" [ 1067.419036] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.427427] env[61728]: DEBUG oslo_vmware.api [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464816, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.457211] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f9f07464-9f43-43fd-8895-2673861747fa] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1067.660698] env[61728]: DEBUG oslo_vmware.api [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464815, 'name': PowerOffVM_Task, 'duration_secs': 0.189105} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.661114] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1067.661325] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1067.661529] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-26136a3d-7581-4a0d-8c9b-be890cc38677 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.724267] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1067.724454] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1067.724587] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleting the datastore file [datastore1] 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.724865] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-68f368ad-9e89-4595-8679-72ddd2ce24b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.735019] env[61728]: DEBUG oslo_vmware.api [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1067.735019] env[61728]: value = "task-464818" [ 1067.735019] env[61728]: _type = "Task" [ 1067.735019] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.739974] env[61728]: DEBUG oslo_vmware.api [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464818, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.879023] env[61728]: DEBUG oslo_concurrency.lockutils [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.791s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.899376] env[61728]: INFO nova.scheduler.client.report [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Deleted allocations for instance 267880ff-0fac-4763-82be-ef8587ee7628 [ 1067.929215] env[61728]: DEBUG oslo_vmware.api [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464816, 'name': ReconfigVM_Task, 'duration_secs': 0.258836} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.929772] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Reconfigured VM instance instance-0000005c to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1067.936056] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f5a92b3-bb0e-4c29-8158-9b64b37c5cbb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.949693] env[61728]: DEBUG oslo_vmware.api [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1067.949693] env[61728]: value = "task-464819" [ 1067.949693] env[61728]: _type = "Task" [ 1067.949693] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.958008] env[61728]: DEBUG oslo_vmware.api [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464819, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.963570] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: fef6cb9e-d8c7-45d3-a7f6-e361ff704006] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1068.112183] env[61728]: DEBUG nova.compute.manager [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1068.149833] env[61728]: DEBUG nova.virt.hardware [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1068.149833] env[61728]: DEBUG nova.virt.hardware [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1068.149833] env[61728]: DEBUG nova.virt.hardware [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1068.149833] env[61728]: DEBUG nova.virt.hardware [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1068.149833] env[61728]: DEBUG nova.virt.hardware [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1068.149833] env[61728]: DEBUG nova.virt.hardware [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1068.149833] env[61728]: DEBUG nova.virt.hardware [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1068.150195] env[61728]: DEBUG nova.virt.hardware [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1068.150341] env[61728]: DEBUG nova.virt.hardware [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1068.150550] env[61728]: DEBUG nova.virt.hardware [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1068.150764] env[61728]: DEBUG nova.virt.hardware [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1068.151699] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1309aef4-4b6b-4ef8-8a3c-a9bc8b4726ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.160279] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e508d41-d42d-4b51-925f-1b12614bf4e0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.244281] env[61728]: DEBUG oslo_vmware.api [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464818, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172469} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.244683] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1068.244944] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1068.245228] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1068.245509] env[61728]: INFO nova.compute.manager [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1068.245847] env[61728]: DEBUG oslo.service.loopingcall [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1068.246565] env[61728]: DEBUG nova.compute.manager [-] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1068.246707] env[61728]: DEBUG nova.network.neutron [-] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1068.373117] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "32a781b0-2db4-4883-ae96-d94c8181902b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.373762] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "32a781b0-2db4-4883-ae96-d94c8181902b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.411970] env[61728]: DEBUG oslo_concurrency.lockutils [None req-613840cb-40e9-4fb0-bf87-a35322683fb0 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "267880ff-0fac-4763-82be-ef8587ee7628" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.303s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.462452] env[61728]: DEBUG oslo_vmware.api [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464819, 'name': ReconfigVM_Task, 'duration_secs': 0.161986} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.462864] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122182', 'volume_id': '9e2549ce-9f21-4a53-af77-978682bfcbbe', 'name': 'volume-9e2549ce-9f21-4a53-af77-978682bfcbbe', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a', 'attached_at': '', 'detached_at': '', 'volume_id': '9e2549ce-9f21-4a53-af77-978682bfcbbe', 'serial': '9e2549ce-9f21-4a53-af77-978682bfcbbe'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1068.467156] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 3415687e-9b8d-42f8-9b18-93951b8ec7da] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1068.660326] env[61728]: DEBUG nova.compute.manager [req-dbce877a-430c-4175-b907-c03034b5df9b req-86d4de21-4fbc-4a7a-b622-a098c96a9b1b service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Received event network-vif-plugged-d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1068.663032] env[61728]: DEBUG oslo_concurrency.lockutils [req-dbce877a-430c-4175-b907-c03034b5df9b req-86d4de21-4fbc-4a7a-b622-a098c96a9b1b service nova] Acquiring lock "532507fa-996d-45c7-bf70-f09c93be79ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.663032] env[61728]: DEBUG oslo_concurrency.lockutils [req-dbce877a-430c-4175-b907-c03034b5df9b req-86d4de21-4fbc-4a7a-b622-a098c96a9b1b service nova] Lock "532507fa-996d-45c7-bf70-f09c93be79ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.663032] env[61728]: DEBUG oslo_concurrency.lockutils [req-dbce877a-430c-4175-b907-c03034b5df9b req-86d4de21-4fbc-4a7a-b622-a098c96a9b1b service nova] Lock "532507fa-996d-45c7-bf70-f09c93be79ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.663032] env[61728]: DEBUG nova.compute.manager [req-dbce877a-430c-4175-b907-c03034b5df9b req-86d4de21-4fbc-4a7a-b622-a098c96a9b1b service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] No waiting events found dispatching network-vif-plugged-d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1068.663953] env[61728]: WARNING nova.compute.manager [req-dbce877a-430c-4175-b907-c03034b5df9b req-86d4de21-4fbc-4a7a-b622-a098c96a9b1b service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Received unexpected event network-vif-plugged-d04bc3a5-d378-4e1e-a145-ec523bcc442a for instance with vm_state building and task_state spawning. [ 1068.699977] env[61728]: DEBUG nova.compute.manager [req-ed282f52-cb73-4e37-9812-85464ddebdf8 req-cf3cd4b0-ec39-4857-b614-54f779ecf61d service nova] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Received event network-vif-deleted-96bfe2c5-12a3-4a70-b2d9-aab955fb6b08 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1068.700090] env[61728]: INFO nova.compute.manager [req-ed282f52-cb73-4e37-9812-85464ddebdf8 req-cf3cd4b0-ec39-4857-b614-54f779ecf61d service nova] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Neutron deleted interface 96bfe2c5-12a3-4a70-b2d9-aab955fb6b08; detaching it from the instance and deleting it from the info cache [ 1068.700245] env[61728]: DEBUG nova.network.neutron [req-ed282f52-cb73-4e37-9812-85464ddebdf8 req-cf3cd4b0-ec39-4857-b614-54f779ecf61d service nova] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.765218] env[61728]: DEBUG nova.network.neutron [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Successfully updated port: d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1068.878608] env[61728]: DEBUG nova.compute.manager [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1068.888183] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "6c154517-68e8-4e70-86eb-cf53f6f812d8" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.888183] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "6c154517-68e8-4e70-86eb-cf53f6f812d8" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.888183] env[61728]: DEBUG nova.compute.manager [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Going to confirm migration 7 {{(pid=61728) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1068.970860] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 40558dd0-69d3-42d6-87cf-8e4ae4ae9dae] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1069.059087] env[61728]: DEBUG nova.objects.instance [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lazy-loading 'flavor' on Instance uuid 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.173124] env[61728]: DEBUG nova.network.neutron [-] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.203602] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eee5b68d-0bd2-4a65-8806-2548d73e1413 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.214761] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a019646-0095-4024-bb31-da2a0aa158dc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.248869] env[61728]: DEBUG nova.compute.manager [req-ed282f52-cb73-4e37-9812-85464ddebdf8 req-cf3cd4b0-ec39-4857-b614-54f779ecf61d service nova] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Detach interface failed, port_id=96bfe2c5-12a3-4a70-b2d9-aab955fb6b08, reason: Instance 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1069.268914] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.269084] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.269245] env[61728]: DEBUG nova.network.neutron [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1069.396928] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.397210] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.398602] env[61728]: INFO nova.compute.claims [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1069.477448] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.477596] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquired lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.477784] env[61728]: DEBUG nova.network.neutron [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1069.477975] env[61728]: DEBUG nova.objects.instance [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lazy-loading 'info_cache' on Instance uuid 6c154517-68e8-4e70-86eb-cf53f6f812d8 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.479641] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 121bef05-ef7b-47ee-b737-15f8b9d91158] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1069.675993] env[61728]: INFO nova.compute.manager [-] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Took 1.43 seconds to deallocate network for instance. [ 1069.783890] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Acquiring lock "5eb3c3fa-d612-43d8-875c-2ee595e29081" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.783979] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Lock "5eb3c3fa-d612-43d8-875c-2ee595e29081" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.807229] env[61728]: DEBUG nova.network.neutron [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1069.950253] env[61728]: DEBUG nova.network.neutron [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updating instance_info_cache with network_info: [{"id": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "address": "fa:16:3e:f1:65:32", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04bc3a5-d3", "ovs_interfaceid": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.985169] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 87540771-c864-4d49-8ad3-b6b559841761] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1070.065924] env[61728]: DEBUG oslo_concurrency.lockutils [None req-52c7622e-af15-495c-a83e-79b6432b804e tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.312s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.183987] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.290027] env[61728]: DEBUG nova.compute.manager [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1070.298197] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.298495] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.004s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.298710] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.298896] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.299088] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.301253] env[61728]: INFO nova.compute.manager [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Terminating instance [ 1070.302818] env[61728]: DEBUG nova.compute.manager [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1070.303017] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1070.303966] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e36b76c0-ea13-4756-a063-efc8cf3aaeca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.313068] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1070.313693] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59cf9eb4-4198-4b04-a730-2858878e4431 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.320980] env[61728]: DEBUG oslo_vmware.api [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1070.320980] env[61728]: value = "task-464820" [ 1070.320980] env[61728]: _type = "Task" [ 1070.320980] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.330047] env[61728]: DEBUG oslo_vmware.api [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464820, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.452889] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.453456] env[61728]: DEBUG nova.compute.manager [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Instance network_info: |[{"id": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "address": "fa:16:3e:f1:65:32", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04bc3a5-d3", "ovs_interfaceid": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1070.454144] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:65:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f856fca-9fb5-41ea-a057-ac4193bd323d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd04bc3a5-d378-4e1e-a145-ec523bcc442a', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1070.461986] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Creating folder: Project (438de94d086e47b99cedc6e4a33471bf). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1070.462334] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a8aaae18-1705-43e4-a8c3-08e89413c840 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.476768] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Created folder: Project (438de94d086e47b99cedc6e4a33471bf) in parent group-v121913. [ 1070.476996] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Creating folder: Instances. Parent ref: group-v122201. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1070.477273] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-84496a37-355b-4412-87a5-9ba139d5fc98 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.488160] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Created folder: Instances in parent group-v122201. [ 1070.488780] env[61728]: DEBUG oslo.service.loopingcall [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1070.488780] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1070.488898] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-770473cf-8903-411c-abe7-79ebb05a0117 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.507604] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: d8657743-30f9-4f8f-aca8-d98f3c557d19] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1070.514852] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1070.514852] env[61728]: value = "task-464823" [ 1070.514852] env[61728]: _type = "Task" [ 1070.514852] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.524287] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464823, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.650532] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f822b4-9857-46eb-8d5a-b40dd2bdf96c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.659769] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-801ca403-1954-4a7f-8fd1-14b06f13df71 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.693366] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d813858-58c5-4a07-8467-ba0cc0f36b80 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.701411] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e1a4c27-adcc-4bfa-a704-05509f94a6cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.714960] env[61728]: DEBUG nova.compute.provider_tree [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.731186] env[61728]: DEBUG nova.network.neutron [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating instance_info_cache with network_info: [{"id": "e7f1c36b-89f0-4858-96a0-3c1875c4bb29", "address": "fa:16:3e:c0:51:65", "network": {"id": "265b2548-0436-44ae-8509-5a20ba336851", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-479521009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08367aaceba548fe93faaedf6371817d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7f1c36b-89", "ovs_interfaceid": "e7f1c36b-89f0-4858-96a0-3c1875c4bb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.788590] env[61728]: DEBUG nova.compute.manager [req-35528445-9943-4e2b-a29f-b23e6e0fb9b2 req-d5eb39e9-471d-48df-822d-a8eb9fa30a22 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Received event network-changed-d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1070.788811] env[61728]: DEBUG nova.compute.manager [req-35528445-9943-4e2b-a29f-b23e6e0fb9b2 req-d5eb39e9-471d-48df-822d-a8eb9fa30a22 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Refreshing instance network info cache due to event network-changed-d04bc3a5-d378-4e1e-a145-ec523bcc442a. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1070.789054] env[61728]: DEBUG oslo_concurrency.lockutils [req-35528445-9943-4e2b-a29f-b23e6e0fb9b2 req-d5eb39e9-471d-48df-822d-a8eb9fa30a22 service nova] Acquiring lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.789215] env[61728]: DEBUG oslo_concurrency.lockutils [req-35528445-9943-4e2b-a29f-b23e6e0fb9b2 req-d5eb39e9-471d-48df-822d-a8eb9fa30a22 service nova] Acquired lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.789384] env[61728]: DEBUG nova.network.neutron [req-35528445-9943-4e2b-a29f-b23e6e0fb9b2 req-d5eb39e9-471d-48df-822d-a8eb9fa30a22 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Refreshing network info cache for port d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1070.816567] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.831784] env[61728]: DEBUG oslo_vmware.api [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464820, 'name': PowerOffVM_Task, 'duration_secs': 0.24299} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.832091] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1070.832269] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1070.832526] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2b7bde4-3552-4d06-89f6-b57f3a28cf9c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.916896] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1070.917135] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1070.917588] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Deleting the datastore file [datastore1] 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1070.917726] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d7652f4-3b8e-4585-b2fa-986a5e2af1e4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.925045] env[61728]: DEBUG oslo_vmware.api [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1070.925045] env[61728]: value = "task-464825" [ 1070.925045] env[61728]: _type = "Task" [ 1070.925045] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.933806] env[61728]: DEBUG oslo_vmware.api [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464825, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.011438] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: a9745dc8-6400-424c-bc10-1181ee128648] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1071.025517] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464823, 'name': CreateVM_Task, 'duration_secs': 0.312997} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.025707] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1071.026464] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.026651] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.026986] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1071.027513] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7488e782-68a6-4c37-a8fb-4588baae103c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.032519] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1071.032519] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bf7cdf-3a1d-460d-662f-088dce0f057b" [ 1071.032519] env[61728]: _type = "Task" [ 1071.032519] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.043023] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bf7cdf-3a1d-460d-662f-088dce0f057b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.056509] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.057140] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.219201] env[61728]: DEBUG nova.scheduler.client.report [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.233840] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Releasing lock "refresh_cache-6c154517-68e8-4e70-86eb-cf53f6f812d8" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.233840] env[61728]: DEBUG nova.objects.instance [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lazy-loading 'migration_context' on Instance uuid 6c154517-68e8-4e70-86eb-cf53f6f812d8 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.435299] env[61728]: DEBUG oslo_vmware.api [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464825, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.512801] env[61728]: DEBUG nova.network.neutron [req-35528445-9943-4e2b-a29f-b23e6e0fb9b2 req-d5eb39e9-471d-48df-822d-a8eb9fa30a22 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updated VIF entry in instance network info cache for port d04bc3a5-d378-4e1e-a145-ec523bcc442a. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1071.513312] env[61728]: DEBUG nova.network.neutron [req-35528445-9943-4e2b-a29f-b23e6e0fb9b2 req-d5eb39e9-471d-48df-822d-a8eb9fa30a22 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updating instance_info_cache with network_info: [{"id": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "address": "fa:16:3e:f1:65:32", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04bc3a5-d3", "ovs_interfaceid": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.514894] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 174607a1-9bc8-4e07-8993-7f0bb0f308e9] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1071.546138] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bf7cdf-3a1d-460d-662f-088dce0f057b, 'name': SearchDatastore_Task, 'duration_secs': 0.009818} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.546138] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.546345] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1071.547028] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.547028] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.547028] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1071.547198] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f70ea59a-dd8e-4041-8588-7ae9e2ae0012 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.556321] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1071.556519] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1071.557317] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c6f81a8-f925-4fef-b1d3-c961d6880aca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.560742] env[61728]: DEBUG nova.compute.manager [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1071.567473] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1071.567473] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5238e041-d724-a433-68cf-9472ec5789ca" [ 1071.567473] env[61728]: _type = "Task" [ 1071.567473] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.575926] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5238e041-d724-a433-68cf-9472ec5789ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.724536] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.327s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.725130] env[61728]: DEBUG nova.compute.manager [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1071.728156] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.544s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.728355] env[61728]: DEBUG nova.objects.instance [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lazy-loading 'resources' on Instance uuid 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.736804] env[61728]: DEBUG nova.objects.base [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Object Instance<6c154517-68e8-4e70-86eb-cf53f6f812d8> lazy-loaded attributes: info_cache,migration_context {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1071.738521] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf028b7-0023-4fcf-8ff1-83073e27dbef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.761609] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ced58181-31c8-447d-b395-ca0375847e92 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.767663] env[61728]: DEBUG oslo_vmware.api [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1071.767663] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529da3db-ba9d-f9cb-6a2b-afece54ca699" [ 1071.767663] env[61728]: _type = "Task" [ 1071.767663] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.776733] env[61728]: DEBUG oslo_vmware.api [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529da3db-ba9d-f9cb-6a2b-afece54ca699, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.937419] env[61728]: DEBUG oslo_vmware.api [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464825, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.543233} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.937745] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1071.937991] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1071.938239] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1071.938464] env[61728]: INFO nova.compute.manager [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Took 1.64 seconds to destroy the instance on the hypervisor. [ 1071.938755] env[61728]: DEBUG oslo.service.loopingcall [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.938999] env[61728]: DEBUG nova.compute.manager [-] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1071.939138] env[61728]: DEBUG nova.network.neutron [-] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1072.017685] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: b444a6cf-3c98-4308-afc1-6e760d30082d] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1072.019959] env[61728]: DEBUG oslo_concurrency.lockutils [req-35528445-9943-4e2b-a29f-b23e6e0fb9b2 req-d5eb39e9-471d-48df-822d-a8eb9fa30a22 service nova] Releasing lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.080564] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5238e041-d724-a433-68cf-9472ec5789ca, 'name': SearchDatastore_Task, 'duration_secs': 0.013808} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.081579] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef8c5c38-1831-4b2c-8352-49a35e587559 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.088203] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.088563] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1072.088563] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a8e6aa-2adc-605f-8ad7-92a952f2f7b2" [ 1072.088563] env[61728]: _type = "Task" [ 1072.088563] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.096575] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a8e6aa-2adc-605f-8ad7-92a952f2f7b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.231033] env[61728]: DEBUG nova.compute.utils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1072.236058] env[61728]: DEBUG nova.compute.manager [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1072.236058] env[61728]: DEBUG nova.network.neutron [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1072.281445] env[61728]: DEBUG oslo_vmware.api [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529da3db-ba9d-f9cb-6a2b-afece54ca699, 'name': SearchDatastore_Task, 'duration_secs': 0.008789} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.284362] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.293744] env[61728]: DEBUG nova.policy [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67d7c6a6191045cf85f790acebaf811e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7f61c87df8b485b9a6370032dd7af37', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1072.484909] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49023d5c-1968-4bc8-bf31-d5beb58c1bee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.493152] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea777005-802b-4cfd-9499-c3dd8f1f1bdb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.526330] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: d44718bd-872b-401c-aa11-f10bea4a35d8] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1072.529033] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f917376-2270-43a3-98ff-5e2a5b139e1d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.538393] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af66d057-fe3a-4fd1-b8cb-f2431574ee4f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.564689] env[61728]: DEBUG nova.compute.provider_tree [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.577786] env[61728]: DEBUG nova.network.neutron [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Successfully created port: c4f8bc0c-ff22-428b-a962-d2d8b3672927 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1072.599250] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a8e6aa-2adc-605f-8ad7-92a952f2f7b2, 'name': SearchDatastore_Task, 'duration_secs': 0.01487} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.599500] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.599910] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 532507fa-996d-45c7-bf70-f09c93be79ed/532507fa-996d-45c7-bf70-f09c93be79ed.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1072.600069] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-424f42b0-a065-4425-b6b9-bc191225113f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.610466] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1072.610466] env[61728]: value = "task-464826" [ 1072.610466] env[61728]: _type = "Task" [ 1072.610466] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.619250] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464826, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.736464] env[61728]: DEBUG nova.compute.manager [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1072.888140] env[61728]: DEBUG nova.network.neutron [-] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.017644] env[61728]: DEBUG nova.compute.manager [req-27250830-64dc-48d1-bef5-e35c13f775ec req-713876b4-e875-4af4-864e-e435e00edd45 service nova] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Received event network-vif-deleted-67077c45-d87f-4d2d-a48b-d23a96d35f8a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1073.035131] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 3ef9bf60-c6a1-4b7a-a375-5397fb871850] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1073.066903] env[61728]: DEBUG nova.scheduler.client.report [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1073.120757] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464826, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478414} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.121084] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 532507fa-996d-45c7-bf70-f09c93be79ed/532507fa-996d-45c7-bf70-f09c93be79ed.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1073.121353] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1073.121641] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da56f113-4fab-4606-be4c-5c3a5ca3bbc2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.128559] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1073.128559] env[61728]: value = "task-464827" [ 1073.128559] env[61728]: _type = "Task" [ 1073.128559] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.136686] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464827, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.393996] env[61728]: INFO nova.compute.manager [-] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Took 1.45 seconds to deallocate network for instance. [ 1073.538428] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 6d55bb01-7855-4d86-ac57-ecae15ea590f] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1073.572392] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.844s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.575470] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.758s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.576248] env[61728]: INFO nova.compute.claims [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1073.595933] env[61728]: INFO nova.scheduler.client.report [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleted allocations for instance 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4 [ 1073.638571] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464827, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08022} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.638873] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1073.639651] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5834a01-8168-4776-82e1-d5f24791ab30 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.663339] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 532507fa-996d-45c7-bf70-f09c93be79ed/532507fa-996d-45c7-bf70-f09c93be79ed.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1073.663871] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b7a9ef8-0974-4c34-b91d-a838cea91d71 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.683993] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1073.683993] env[61728]: value = "task-464828" [ 1073.683993] env[61728]: _type = "Task" [ 1073.683993] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.691953] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464828, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.746105] env[61728]: DEBUG nova.compute.manager [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1073.772293] env[61728]: DEBUG nova.virt.hardware [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1073.772596] env[61728]: DEBUG nova.virt.hardware [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1073.772801] env[61728]: DEBUG nova.virt.hardware [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1073.773034] env[61728]: DEBUG nova.virt.hardware [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1073.773232] env[61728]: DEBUG nova.virt.hardware [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1073.773403] env[61728]: DEBUG nova.virt.hardware [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1073.773656] env[61728]: DEBUG nova.virt.hardware [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1073.773868] env[61728]: DEBUG nova.virt.hardware [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1073.774076] env[61728]: DEBUG nova.virt.hardware [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1073.774266] env[61728]: DEBUG nova.virt.hardware [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1073.774496] env[61728]: DEBUG nova.virt.hardware [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1073.775440] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d19ef812-d771-4e49-8ef4-2f78cf81d2b4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.783659] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe7570b-d04f-4925-a897-c1e20352ef37 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.900790] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.043576] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 06109957-2c3b-404e-a43e-dd34ece39096] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1074.103842] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f5cc6aaf-9e0f-48d0-b9e0-208e34fd2cd8 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.992s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.194469] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464828, 'name': ReconfigVM_Task, 'duration_secs': 0.330888} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.194745] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 532507fa-996d-45c7-bf70-f09c93be79ed/532507fa-996d-45c7-bf70-f09c93be79ed.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1074.195554] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-db4954b8-243a-43db-b112-a3e7f36bc16b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.201865] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1074.201865] env[61728]: value = "task-464829" [ 1074.201865] env[61728]: _type = "Task" [ 1074.201865] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.209265] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464829, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.286267] env[61728]: DEBUG nova.network.neutron [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Successfully updated port: c4f8bc0c-ff22-428b-a962-d2d8b3672927 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1074.427433] env[61728]: DEBUG nova.compute.manager [req-834c69ac-9d76-41bb-b4e1-da8c0195505b req-dd08f5fe-e844-4d72-a757-60325fd50129 service nova] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Received event network-vif-plugged-c4f8bc0c-ff22-428b-a962-d2d8b3672927 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1074.427672] env[61728]: DEBUG oslo_concurrency.lockutils [req-834c69ac-9d76-41bb-b4e1-da8c0195505b req-dd08f5fe-e844-4d72-a757-60325fd50129 service nova] Acquiring lock "32a781b0-2db4-4883-ae96-d94c8181902b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.427885] env[61728]: DEBUG oslo_concurrency.lockutils [req-834c69ac-9d76-41bb-b4e1-da8c0195505b req-dd08f5fe-e844-4d72-a757-60325fd50129 service nova] Lock "32a781b0-2db4-4883-ae96-d94c8181902b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.428081] env[61728]: DEBUG oslo_concurrency.lockutils [req-834c69ac-9d76-41bb-b4e1-da8c0195505b req-dd08f5fe-e844-4d72-a757-60325fd50129 service nova] Lock "32a781b0-2db4-4883-ae96-d94c8181902b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.428262] env[61728]: DEBUG nova.compute.manager [req-834c69ac-9d76-41bb-b4e1-da8c0195505b req-dd08f5fe-e844-4d72-a757-60325fd50129 service nova] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] No waiting events found dispatching network-vif-plugged-c4f8bc0c-ff22-428b-a962-d2d8b3672927 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1074.428431] env[61728]: WARNING nova.compute.manager [req-834c69ac-9d76-41bb-b4e1-da8c0195505b req-dd08f5fe-e844-4d72-a757-60325fd50129 service nova] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Received unexpected event network-vif-plugged-c4f8bc0c-ff22-428b-a962-d2d8b3672927 for instance with vm_state building and task_state spawning. [ 1074.547263] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: b1d62341-1cf1-49c4-bc4b-e5e0261aa5d4] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1074.712853] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464829, 'name': Rename_Task, 'duration_secs': 0.157075} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.715166] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1074.715639] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-343a5890-ae61-4cd3-98f1-af2c3dd66504 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.721968] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1074.721968] env[61728]: value = "task-464830" [ 1074.721968] env[61728]: _type = "Task" [ 1074.721968] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.731890] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464830, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.767494] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8672946-c269-4783-8b5e-a24feb717b2f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.775073] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6c1113-ab22-4e7b-9df2-457c061f912e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.806178] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-32a781b0-2db4-4883-ae96-d94c8181902b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.806332] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-32a781b0-2db4-4883-ae96-d94c8181902b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.806484] env[61728]: DEBUG nova.network.neutron [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1074.808481] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752159bc-0332-4f6a-bad4-8cc3c5977a53 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.816895] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a982e04-83ee-4f81-8898-6a3d20e4d474 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.831700] env[61728]: DEBUG nova.compute.provider_tree [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.035356] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "8ea37e93-16fc-4c60-9949-17656218d46a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.035600] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "8ea37e93-16fc-4c60-9949-17656218d46a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.049973] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: a716cf5a-4482-45cb-96ec-b8c38bc3e742] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1075.232512] env[61728]: DEBUG oslo_vmware.api [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464830, 'name': PowerOnVM_Task, 'duration_secs': 0.4978} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.232882] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1075.233154] env[61728]: INFO nova.compute.manager [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Took 7.12 seconds to spawn the instance on the hypervisor. [ 1075.233361] env[61728]: DEBUG nova.compute.manager [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1075.234173] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a97196-12cc-44f0-bebd-c3ff8803f4d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.335328] env[61728]: DEBUG nova.scheduler.client.report [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1075.343103] env[61728]: DEBUG nova.network.neutron [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1075.480349] env[61728]: DEBUG nova.network.neutron [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Updating instance_info_cache with network_info: [{"id": "c4f8bc0c-ff22-428b-a962-d2d8b3672927", "address": "fa:16:3e:6a:84:a6", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4f8bc0c-ff", "ovs_interfaceid": "c4f8bc0c-ff22-428b-a962-d2d8b3672927", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.538018] env[61728]: DEBUG nova.compute.manager [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1075.553453] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f28348d2-c062-497a-b374-521df51054ee] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1075.755089] env[61728]: INFO nova.compute.manager [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Took 13.48 seconds to build instance. [ 1075.841558] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.267s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.842196] env[61728]: DEBUG nova.compute.manager [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1075.844895] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.757s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.846781] env[61728]: INFO nova.compute.claims [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1075.983332] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-32a781b0-2db4-4883-ae96-d94c8181902b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.983804] env[61728]: DEBUG nova.compute.manager [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Instance network_info: |[{"id": "c4f8bc0c-ff22-428b-a962-d2d8b3672927", "address": "fa:16:3e:6a:84:a6", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4f8bc0c-ff", "ovs_interfaceid": "c4f8bc0c-ff22-428b-a962-d2d8b3672927", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1075.984785] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:84:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b80dd748-3d7e-4a23-a38d-9e79a3881452', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c4f8bc0c-ff22-428b-a962-d2d8b3672927', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1075.997049] env[61728]: DEBUG oslo.service.loopingcall [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.997411] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1075.997747] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bffd202a-6fab-4425-aedf-e19d2e94fc94 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.028454] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1076.028454] env[61728]: value = "task-464831" [ 1076.028454] env[61728]: _type = "Task" [ 1076.028454] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.036831] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464831, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.056424] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: a5c7e3ff-6829-4528-9279-e865a0eb8512] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1076.069764] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.256749] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a4269d4a-9c50-4b78-95ac-88e862dce87c tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "532507fa-996d-45c7-bf70-f09c93be79ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.992s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.351149] env[61728]: DEBUG nova.compute.utils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1076.354982] env[61728]: DEBUG nova.compute.manager [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1076.355196] env[61728]: DEBUG nova.network.neutron [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1076.409122] env[61728]: DEBUG nova.policy [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb0626acb0874cd9b39e94ee6c50be46', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c3050bce6074ca9865f5bdea08255af', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1076.459214] env[61728]: DEBUG nova.compute.manager [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Received event network-changed-c4f8bc0c-ff22-428b-a962-d2d8b3672927 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1076.459528] env[61728]: DEBUG nova.compute.manager [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Refreshing instance network info cache due to event network-changed-c4f8bc0c-ff22-428b-a962-d2d8b3672927. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1076.459766] env[61728]: DEBUG oslo_concurrency.lockutils [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] Acquiring lock "refresh_cache-32a781b0-2db4-4883-ae96-d94c8181902b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.459877] env[61728]: DEBUG oslo_concurrency.lockutils [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] Acquired lock "refresh_cache-32a781b0-2db4-4883-ae96-d94c8181902b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.460217] env[61728]: DEBUG nova.network.neutron [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Refreshing network info cache for port c4f8bc0c-ff22-428b-a962-d2d8b3672927 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1076.538449] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464831, 'name': CreateVM_Task, 'duration_secs': 0.326038} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.538628] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1076.539323] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.539516] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.539909] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1076.540058] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec1ef063-0a43-4ff7-9b8f-76cf866a7b33 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.544508] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1076.544508] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521ed005-3fb2-8b46-a466-e3425bc65a49" [ 1076.544508] env[61728]: _type = "Task" [ 1076.544508] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.551737] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521ed005-3fb2-8b46-a466-e3425bc65a49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.560224] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 06e765e1-85d9-4f6c-8ab5-81735f08cb66] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1076.737496] env[61728]: DEBUG nova.network.neutron [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Successfully created port: e8e000ef-5e57-43c4-9e1c-6d296929f225 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1076.856365] env[61728]: DEBUG nova.compute.manager [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1077.058463] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521ed005-3fb2-8b46-a466-e3425bc65a49, 'name': SearchDatastore_Task, 'duration_secs': 0.010093} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.058769] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.059020] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1077.059268] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.059415] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.059610] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1077.059936] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bfd3b5bf-ab95-4e37-8aaa-72b1f4821cdc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.063725] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: b63fa3df-317a-46d4-b8e9-74e9e287efde] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1077.074472] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1077.074648] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1077.075632] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2b23e10-afde-4ff7-bf6f-39b71a7d9bdd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.079517] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a972c403-e534-4929-91c0-0fbfa3109b3c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.084744] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1077.084744] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528ea582-2957-c6af-d214-3497ac39a0a1" [ 1077.084744] env[61728]: _type = "Task" [ 1077.084744] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.089800] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f228bfb-2de5-4147-b470-d624bb533bd6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.101276] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528ea582-2957-c6af-d214-3497ac39a0a1, 'name': SearchDatastore_Task, 'duration_secs': 0.008188} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.127599] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d81a149-cb23-48d0-8703-8895c7becf0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.130197] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d053963d-0ca2-495b-84d4-3ef314eb4b05 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.135714] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1077.135714] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522e00c5-62d3-0d07-c78b-f89410a23fb9" [ 1077.135714] env[61728]: _type = "Task" [ 1077.135714] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.141655] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89bf4e0b-e847-4c9a-84c4-c72e6206ff78 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.157705] env[61728]: DEBUG nova.compute.provider_tree [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.158878] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522e00c5-62d3-0d07-c78b-f89410a23fb9, 'name': SearchDatastore_Task, 'duration_secs': 0.008918} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.159138] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.159386] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 32a781b0-2db4-4883-ae96-d94c8181902b/32a781b0-2db4-4883-ae96-d94c8181902b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1077.159626] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7b24f615-d217-42e3-8af6-4b9a5a9e67e2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.165964] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1077.165964] env[61728]: value = "task-464832" [ 1077.165964] env[61728]: _type = "Task" [ 1077.165964] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.177329] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464832, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.205366] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.205608] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.207747] env[61728]: DEBUG nova.network.neutron [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Updated VIF entry in instance network info cache for port c4f8bc0c-ff22-428b-a962-d2d8b3672927. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1077.208079] env[61728]: DEBUG nova.network.neutron [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Updating instance_info_cache with network_info: [{"id": "c4f8bc0c-ff22-428b-a962-d2d8b3672927", "address": "fa:16:3e:6a:84:a6", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4f8bc0c-ff", "ovs_interfaceid": "c4f8bc0c-ff22-428b-a962-d2d8b3672927", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.569639] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 39eaa6f4-df3f-4928-a1f1-8c861a392e29] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1077.663312] env[61728]: DEBUG nova.scheduler.client.report [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1077.675607] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464832, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462247} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.677432] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 32a781b0-2db4-4883-ae96-d94c8181902b/32a781b0-2db4-4883-ae96-d94c8181902b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1077.677432] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1077.677432] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f42910bc-d82a-4c80-b9d4-7ea239a56caa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.682945] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1077.682945] env[61728]: value = "task-464833" [ 1077.682945] env[61728]: _type = "Task" [ 1077.682945] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.691333] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464833, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.710724] env[61728]: DEBUG oslo_concurrency.lockutils [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] Releasing lock "refresh_cache-32a781b0-2db4-4883-ae96-d94c8181902b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.711190] env[61728]: DEBUG nova.compute.manager [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Received event network-changed-d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1077.711507] env[61728]: DEBUG nova.compute.manager [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Refreshing instance network info cache due to event network-changed-d04bc3a5-d378-4e1e-a145-ec523bcc442a. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1077.711863] env[61728]: DEBUG oslo_concurrency.lockutils [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] Acquiring lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.712100] env[61728]: DEBUG oslo_concurrency.lockutils [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] Acquired lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.712303] env[61728]: DEBUG nova.network.neutron [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Refreshing network info cache for port d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1077.714434] env[61728]: DEBUG nova.compute.utils [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1077.869301] env[61728]: DEBUG nova.compute.manager [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1077.895562] env[61728]: DEBUG nova.virt.hardware [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1077.895824] env[61728]: DEBUG nova.virt.hardware [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1077.896014] env[61728]: DEBUG nova.virt.hardware [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1077.896229] env[61728]: DEBUG nova.virt.hardware [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1077.896388] env[61728]: DEBUG nova.virt.hardware [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1077.896544] env[61728]: DEBUG nova.virt.hardware [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1077.896757] env[61728]: DEBUG nova.virt.hardware [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1077.896938] env[61728]: DEBUG nova.virt.hardware [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1077.897128] env[61728]: DEBUG nova.virt.hardware [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1077.897301] env[61728]: DEBUG nova.virt.hardware [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1077.897506] env[61728]: DEBUG nova.virt.hardware [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1077.898404] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67aa03e8-9215-455b-918a-0255d2e1fcd0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.906825] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf60ff1-f441-4237-a602-d07c1abb3ee0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.073324] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: dc8a587f-3dc8-43b3-b986-2096e3c4e1f1] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1078.171128] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.326s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.171681] env[61728]: DEBUG nova.compute.manager [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1078.174429] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 5.890s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.193328] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464833, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071436} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.194041] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1078.194369] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0aadcb4-d290-42c3-93e8-e4eaed1352a0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.219482] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] 32a781b0-2db4-4883-ae96-d94c8181902b/32a781b0-2db4-4883-ae96-d94c8181902b.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1078.222198] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.016s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.222664] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9c51f6d-e485-48c4-af10-ccf6dd5a6b12 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.243349] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1078.243349] env[61728]: value = "task-464834" [ 1078.243349] env[61728]: _type = "Task" [ 1078.243349] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.251483] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464834, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.444095] env[61728]: DEBUG nova.network.neutron [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updated VIF entry in instance network info cache for port d04bc3a5-d378-4e1e-a145-ec523bcc442a. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1078.444571] env[61728]: DEBUG nova.network.neutron [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updating instance_info_cache with network_info: [{"id": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "address": "fa:16:3e:f1:65:32", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04bc3a5-d3", "ovs_interfaceid": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.576987] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f2692e22-6c14-4bb3-a080-607f0731105d] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1078.681212] env[61728]: DEBUG nova.compute.utils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1078.682878] env[61728]: DEBUG nova.compute.manager [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1078.683070] env[61728]: DEBUG nova.network.neutron [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1078.724948] env[61728]: DEBUG nova.policy [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '70856f8255c940ac979244efb02c43b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '276942c8cf8a42729a541de096f69a0c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1078.756336] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464834, 'name': ReconfigVM_Task, 'duration_secs': 0.293542} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.756642] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Reconfigured VM instance instance-00000070 to attach disk [datastore1] 32a781b0-2db4-4883-ae96-d94c8181902b/32a781b0-2db4-4883-ae96-d94c8181902b.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1078.757335] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8bcc465f-b1a5-4a08-98e3-92251e33170e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.763850] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1078.763850] env[61728]: value = "task-464835" [ 1078.763850] env[61728]: _type = "Task" [ 1078.763850] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.771756] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464835, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.908742] env[61728]: DEBUG nova.network.neutron [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Successfully updated port: e8e000ef-5e57-43c4-9e1c-6d296929f225 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1078.916383] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7c6c56-7088-4a4c-accd-09953655f343 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.927327] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ddf878-00aa-466a-8b6c-11966a7abef4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.959680] env[61728]: DEBUG oslo_concurrency.lockutils [req-f3891825-796e-4acb-b929-1185dd7f6253 req-068ade9e-a2fb-43a8-92d5-4087c8389d55 service nova] Releasing lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.960831] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1231b8a-027c-4b94-9dca-8edabf9fa0ea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.969169] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a6a9a2-8bdc-4e70-8d71-16381f49a2ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.982484] env[61728]: DEBUG nova.compute.provider_tree [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.061021] env[61728]: DEBUG nova.network.neutron [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Successfully created port: 4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1079.080518] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: c84bfcd0-b145-4675-8b0a-5e8f94f65098] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1079.186301] env[61728]: DEBUG nova.compute.manager [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1079.274290] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464835, 'name': Rename_Task, 'duration_secs': 0.145673} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.274646] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1079.274939] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cd0cebde-d51f-4ef6-b0a0-291de14301b7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.281701] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1079.281701] env[61728]: value = "task-464836" [ 1079.281701] env[61728]: _type = "Task" [ 1079.281701] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.285584] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.285737] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.285953] env[61728]: INFO nova.compute.manager [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Attaching volume 76406f6c-49da-4712-8faa-8af3a2f760a1 to /dev/sdb [ 1079.292293] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464836, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.325295] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab28cce-b550-4424-9ed3-e2c72a3f71c8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.332633] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e53268d-39f9-4552-9b15-0d1134132e63 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.346349] env[61728]: DEBUG nova.virt.block_device [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Updating existing volume attachment record: 49db116d-b36a-4f27-a2b3-6785627112d4 {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1079.364896] env[61728]: DEBUG nova.compute.manager [req-040d2a5e-6914-47ac-9f1e-d988d68f4a7a req-5ba8588f-c367-4e52-9bb1-35914ea48391 service nova] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Received event network-vif-plugged-e8e000ef-5e57-43c4-9e1c-6d296929f225 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1079.365096] env[61728]: DEBUG oslo_concurrency.lockutils [req-040d2a5e-6914-47ac-9f1e-d988d68f4a7a req-5ba8588f-c367-4e52-9bb1-35914ea48391 service nova] Acquiring lock "5eb3c3fa-d612-43d8-875c-2ee595e29081-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.365349] env[61728]: DEBUG oslo_concurrency.lockutils [req-040d2a5e-6914-47ac-9f1e-d988d68f4a7a req-5ba8588f-c367-4e52-9bb1-35914ea48391 service nova] Lock "5eb3c3fa-d612-43d8-875c-2ee595e29081-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.365476] env[61728]: DEBUG oslo_concurrency.lockutils [req-040d2a5e-6914-47ac-9f1e-d988d68f4a7a req-5ba8588f-c367-4e52-9bb1-35914ea48391 service nova] Lock "5eb3c3fa-d612-43d8-875c-2ee595e29081-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.365649] env[61728]: DEBUG nova.compute.manager [req-040d2a5e-6914-47ac-9f1e-d988d68f4a7a req-5ba8588f-c367-4e52-9bb1-35914ea48391 service nova] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] No waiting events found dispatching network-vif-plugged-e8e000ef-5e57-43c4-9e1c-6d296929f225 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1079.365815] env[61728]: WARNING nova.compute.manager [req-040d2a5e-6914-47ac-9f1e-d988d68f4a7a req-5ba8588f-c367-4e52-9bb1-35914ea48391 service nova] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Received unexpected event network-vif-plugged-e8e000ef-5e57-43c4-9e1c-6d296929f225 for instance with vm_state building and task_state spawning. [ 1079.381837] env[61728]: DEBUG nova.compute.manager [req-60b489db-26d3-43ff-8a81-238a981645d6 req-a064cebe-c8cc-4cc4-8c6e-c441f7ae2ab7 service nova] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Received event network-changed-e8e000ef-5e57-43c4-9e1c-6d296929f225 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1079.382091] env[61728]: DEBUG nova.compute.manager [req-60b489db-26d3-43ff-8a81-238a981645d6 req-a064cebe-c8cc-4cc4-8c6e-c441f7ae2ab7 service nova] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Refreshing instance network info cache due to event network-changed-e8e000ef-5e57-43c4-9e1c-6d296929f225. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1079.382366] env[61728]: DEBUG oslo_concurrency.lockutils [req-60b489db-26d3-43ff-8a81-238a981645d6 req-a064cebe-c8cc-4cc4-8c6e-c441f7ae2ab7 service nova] Acquiring lock "refresh_cache-5eb3c3fa-d612-43d8-875c-2ee595e29081" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.382454] env[61728]: DEBUG oslo_concurrency.lockutils [req-60b489db-26d3-43ff-8a81-238a981645d6 req-a064cebe-c8cc-4cc4-8c6e-c441f7ae2ab7 service nova] Acquired lock "refresh_cache-5eb3c3fa-d612-43d8-875c-2ee595e29081" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.382601] env[61728]: DEBUG nova.network.neutron [req-60b489db-26d3-43ff-8a81-238a981645d6 req-a064cebe-c8cc-4cc4-8c6e-c441f7ae2ab7 service nova] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Refreshing network info cache for port e8e000ef-5e57-43c4-9e1c-6d296929f225 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1079.411429] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Acquiring lock "refresh_cache-5eb3c3fa-d612-43d8-875c-2ee595e29081" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.487105] env[61728]: DEBUG nova.scheduler.client.report [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.584167] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 6df7b619-8cc4-4dd9-8596-22dc83234a8e] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1079.791677] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464836, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.912828] env[61728]: DEBUG nova.network.neutron [req-60b489db-26d3-43ff-8a81-238a981645d6 req-a064cebe-c8cc-4cc4-8c6e-c441f7ae2ab7 service nova] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1079.982401] env[61728]: DEBUG nova.network.neutron [req-60b489db-26d3-43ff-8a81-238a981645d6 req-a064cebe-c8cc-4cc4-8c6e-c441f7ae2ab7 service nova] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.086797] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: bbf07a5a-42e5-4ac7-8163-3e399dfa9ef5] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1080.195892] env[61728]: DEBUG nova.compute.manager [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1080.222678] env[61728]: DEBUG nova.virt.hardware [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1080.222948] env[61728]: DEBUG nova.virt.hardware [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1080.223134] env[61728]: DEBUG nova.virt.hardware [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1080.223328] env[61728]: DEBUG nova.virt.hardware [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1080.223481] env[61728]: DEBUG nova.virt.hardware [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1080.223633] env[61728]: DEBUG nova.virt.hardware [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1080.223851] env[61728]: DEBUG nova.virt.hardware [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1080.224057] env[61728]: DEBUG nova.virt.hardware [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1080.224253] env[61728]: DEBUG nova.virt.hardware [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1080.224426] env[61728]: DEBUG nova.virt.hardware [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1080.224608] env[61728]: DEBUG nova.virt.hardware [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.225506] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08dcd004-34e2-4109-ad73-acd3172ec06a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.233786] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515b138e-fc40-4602-80fd-a0f594e54ae1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.293236] env[61728]: DEBUG oslo_vmware.api [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464836, 'name': PowerOnVM_Task, 'duration_secs': 0.582257} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.293548] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1080.293755] env[61728]: INFO nova.compute.manager [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Took 6.55 seconds to spawn the instance on the hypervisor. [ 1080.293939] env[61728]: DEBUG nova.compute.manager [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1080.294738] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c6434a2-0691-416a-928d-0042a77147ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.485216] env[61728]: DEBUG oslo_concurrency.lockutils [req-60b489db-26d3-43ff-8a81-238a981645d6 req-a064cebe-c8cc-4cc4-8c6e-c441f7ae2ab7 service nova] Releasing lock "refresh_cache-5eb3c3fa-d612-43d8-875c-2ee595e29081" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.485620] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Acquired lock "refresh_cache-5eb3c3fa-d612-43d8-875c-2ee595e29081" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.485809] env[61728]: DEBUG nova.network.neutron [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1080.490192] env[61728]: DEBUG nova.compute.manager [req-3808e5df-9097-415e-aae3-4ee69058f6b1 req-482a4529-c572-47ac-b93b-9eaca9e4cf6f service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Received event network-vif-plugged-4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1080.490412] env[61728]: DEBUG oslo_concurrency.lockutils [req-3808e5df-9097-415e-aae3-4ee69058f6b1 req-482a4529-c572-47ac-b93b-9eaca9e4cf6f service nova] Acquiring lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.490624] env[61728]: DEBUG oslo_concurrency.lockutils [req-3808e5df-9097-415e-aae3-4ee69058f6b1 req-482a4529-c572-47ac-b93b-9eaca9e4cf6f service nova] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.490797] env[61728]: DEBUG oslo_concurrency.lockutils [req-3808e5df-9097-415e-aae3-4ee69058f6b1 req-482a4529-c572-47ac-b93b-9eaca9e4cf6f service nova] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.490976] env[61728]: DEBUG nova.compute.manager [req-3808e5df-9097-415e-aae3-4ee69058f6b1 req-482a4529-c572-47ac-b93b-9eaca9e4cf6f service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] No waiting events found dispatching network-vif-plugged-4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1080.491314] env[61728]: WARNING nova.compute.manager [req-3808e5df-9097-415e-aae3-4ee69058f6b1 req-482a4529-c572-47ac-b93b-9eaca9e4cf6f service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Received unexpected event network-vif-plugged-4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd for instance with vm_state building and task_state spawning. [ 1080.498222] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.324s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.500787] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.600s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.501087] env[61728]: DEBUG nova.objects.instance [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lazy-loading 'resources' on Instance uuid 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1080.524564] env[61728]: DEBUG nova.network.neutron [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Successfully updated port: 4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1080.591034] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 3bfefcfc-db97-4a9d-86cb-9fb1d8158863] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1080.811580] env[61728]: INFO nova.compute.manager [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Took 11.43 seconds to build instance. [ 1081.020658] env[61728]: DEBUG nova.network.neutron [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1081.029482] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "refresh_cache-c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.029482] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired lock "refresh_cache-c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.029482] env[61728]: DEBUG nova.network.neutron [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1081.061865] env[61728]: INFO nova.scheduler.client.report [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleted allocation for migration 3f7c1af7-63cb-44c6-b2d2-a0146aa543d0 [ 1081.093270] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 1d9114cb-a0af-4874-962f-27237b3c89cc] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1081.185484] env[61728]: DEBUG nova.network.neutron [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Updating instance_info_cache with network_info: [{"id": "e8e000ef-5e57-43c4-9e1c-6d296929f225", "address": "fa:16:3e:ff:e1:ab", "network": {"id": "db5d6fb7-ff65-41a2-b629-7cb0e0aaa55f", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-861227247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c3050bce6074ca9865f5bdea08255af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "842f738f-eaa4-4444-a9bf-90d2b533184c", "external-id": "nsx-vlan-transportzone-460", "segmentation_id": 460, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8e000ef-5e", "ovs_interfaceid": "e8e000ef-5e57-43c4-9e1c-6d296929f225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.218860] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8202a2c9-9fe6-4b07-b47f-5dd93e51ed82 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.228472] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e1b920-295a-4cfe-89f0-421b582e6aaf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.257151] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9815fb-ad41-45ee-9b82-753f6cf8cbdf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.264635] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf9cb0b-0763-4156-8df5-e40ca18abe25 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.277726] env[61728]: DEBUG nova.compute.provider_tree [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.313804] env[61728]: DEBUG oslo_concurrency.lockutils [None req-811efd68-132c-4038-a261-bdb974cb9f0a tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "32a781b0-2db4-4883-ae96-d94c8181902b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.940s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.386031] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.386167] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.571629] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "6c154517-68e8-4e70-86eb-cf53f6f812d8" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 12.684s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.596338] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 0fb1192e-99f1-4469-b196-60df7eab8185] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1081.690277] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Releasing lock "refresh_cache-5eb3c3fa-d612-43d8-875c-2ee595e29081" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.690614] env[61728]: DEBUG nova.compute.manager [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Instance network_info: |[{"id": "e8e000ef-5e57-43c4-9e1c-6d296929f225", "address": "fa:16:3e:ff:e1:ab", "network": {"id": "db5d6fb7-ff65-41a2-b629-7cb0e0aaa55f", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-861227247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c3050bce6074ca9865f5bdea08255af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "842f738f-eaa4-4444-a9bf-90d2b533184c", "external-id": "nsx-vlan-transportzone-460", "segmentation_id": 460, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8e000ef-5e", "ovs_interfaceid": "e8e000ef-5e57-43c4-9e1c-6d296929f225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1081.691061] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:e1:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '842f738f-eaa4-4444-a9bf-90d2b533184c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e8e000ef-5e57-43c4-9e1c-6d296929f225', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1081.698621] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Creating folder: Project (3c3050bce6074ca9865f5bdea08255af). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1081.698920] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-afaf3e91-a7c4-4cc0-9b8d-9a41e5431e31 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.709925] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Created folder: Project (3c3050bce6074ca9865f5bdea08255af) in parent group-v121913. [ 1081.710139] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Creating folder: Instances. Parent ref: group-v122207. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1081.710366] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f7c2fd1-2d47-4d1a-b4c3-ace41058c9fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.719889] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Created folder: Instances in parent group-v122207. [ 1081.720133] env[61728]: DEBUG oslo.service.loopingcall [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1081.720326] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1081.720550] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cb6e6d49-9169-450e-9217-866d8253a6f9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.734559] env[61728]: DEBUG nova.network.neutron [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1081.740514] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1081.740514] env[61728]: value = "task-464843" [ 1081.740514] env[61728]: _type = "Task" [ 1081.740514] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.747796] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464843, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.780202] env[61728]: DEBUG nova.scheduler.client.report [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1081.864462] env[61728]: DEBUG nova.network.neutron [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Updating instance_info_cache with network_info: [{"id": "4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd", "address": "fa:16:3e:18:23:bd", "network": {"id": "22ad4b68-2895-4651-803f-44b8031a6292", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-489227259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "276942c8cf8a42729a541de096f69a0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c25f3d2-3e", "ovs_interfaceid": "4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.889814] env[61728]: DEBUG nova.compute.utils [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1082.099559] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: e3116a87-ce12-4ec8-b6fa-70ec2127ebb0] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1082.250935] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464843, 'name': CreateVM_Task, 'duration_secs': 0.309455} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.251204] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1082.252096] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.252367] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.252679] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1082.253240] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-531b41c6-7662-4498-9dbb-cd77ae16feaf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.257481] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Waiting for the task: (returnval){ [ 1082.257481] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520f0bfb-a67b-c3bc-2e40-d5f5bf3fa137" [ 1082.257481] env[61728]: _type = "Task" [ 1082.257481] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.264871] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520f0bfb-a67b-c3bc-2e40-d5f5bf3fa137, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.284743] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.784s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.287144] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.217s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.288761] env[61728]: INFO nova.compute.claims [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1082.308163] env[61728]: INFO nova.scheduler.client.report [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Deleted allocations for instance 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a [ 1082.367235] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Releasing lock "refresh_cache-c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.367527] env[61728]: DEBUG nova.compute.manager [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Instance network_info: |[{"id": "4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd", "address": "fa:16:3e:18:23:bd", "network": {"id": "22ad4b68-2895-4651-803f-44b8031a6292", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-489227259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "276942c8cf8a42729a541de096f69a0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c25f3d2-3e", "ovs_interfaceid": "4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1082.368175] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:23:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0248a27a-1d7f-4195-987b-06bfc8467347', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1082.375993] env[61728]: DEBUG oslo.service.loopingcall [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1082.376510] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1082.377124] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-47ee8811-4883-4613-8e0e-497f78a3fdcc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.392062] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.397797] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1082.397797] env[61728]: value = "task-464844" [ 1082.397797] env[61728]: _type = "Task" [ 1082.397797] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.405471] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464844, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.602869] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: c2e5a2ea-1d08-4acd-99b1-2ccd55b4c4df] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1082.768433] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520f0bfb-a67b-c3bc-2e40-d5f5bf3fa137, 'name': SearchDatastore_Task, 'duration_secs': 0.007927} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.768754] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.769014] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1082.769648] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.769648] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.769648] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1082.769913] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0b2c781-87c6-48d5-b0db-dafafc74a3eb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.778186] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1082.778359] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1082.779143] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59f05f76-f067-45aa-b341-a4793800730a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.784602] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Waiting for the task: (returnval){ [ 1082.784602] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52807baa-c014-fd13-7dfb-a83088c68b10" [ 1082.784602] env[61728]: _type = "Task" [ 1082.784602] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.794874] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52807baa-c014-fd13-7dfb-a83088c68b10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.815842] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a5855599-e20a-46d1-ad67-bbdc770c281f tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.517s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.908500] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464844, 'name': CreateVM_Task, 'duration_secs': 0.464071} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.908741] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1082.909629] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.909891] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.910263] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1082.910559] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b660c6d-cd24-4bc3-9369-4ac59e8c1670 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.915090] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1082.915090] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526296c3-a74d-f9a7-c82e-4c9c4834fe20" [ 1082.915090] env[61728]: _type = "Task" [ 1082.915090] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.918343] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "6c154517-68e8-4e70-86eb-cf53f6f812d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.918590] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "6c154517-68e8-4e70-86eb-cf53f6f812d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.918849] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "6c154517-68e8-4e70-86eb-cf53f6f812d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.919045] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "6c154517-68e8-4e70-86eb-cf53f6f812d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.919471] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "6c154517-68e8-4e70-86eb-cf53f6f812d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.925535] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526296c3-a74d-f9a7-c82e-4c9c4834fe20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.926094] env[61728]: INFO nova.compute.manager [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Terminating instance [ 1082.928542] env[61728]: DEBUG nova.compute.manager [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1082.928774] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1082.929633] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203803a7-8311-4edb-95c4-f3e217be251b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.936467] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1082.936717] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f51c2e8d-5977-4a8a-a456-819ef26ef34e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.942748] env[61728]: DEBUG oslo_vmware.api [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1082.942748] env[61728]: value = "task-464845" [ 1082.942748] env[61728]: _type = "Task" [ 1082.942748] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.951857] env[61728]: DEBUG oslo_vmware.api [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464845, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.984188] env[61728]: DEBUG nova.compute.manager [req-fa4e724b-825f-4525-987b-c4035cb68c06 req-704375c3-4acc-406e-a2b6-766432af7023 service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Received event network-changed-4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1082.984188] env[61728]: DEBUG nova.compute.manager [req-fa4e724b-825f-4525-987b-c4035cb68c06 req-704375c3-4acc-406e-a2b6-766432af7023 service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Refreshing instance network info cache due to event network-changed-4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1082.984188] env[61728]: DEBUG oslo_concurrency.lockutils [req-fa4e724b-825f-4525-987b-c4035cb68c06 req-704375c3-4acc-406e-a2b6-766432af7023 service nova] Acquiring lock "refresh_cache-c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.984188] env[61728]: DEBUG oslo_concurrency.lockutils [req-fa4e724b-825f-4525-987b-c4035cb68c06 req-704375c3-4acc-406e-a2b6-766432af7023 service nova] Acquired lock "refresh_cache-c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.984188] env[61728]: DEBUG nova.network.neutron [req-fa4e724b-825f-4525-987b-c4035cb68c06 req-704375c3-4acc-406e-a2b6-766432af7023 service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Refreshing network info cache for port 4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1083.106301] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 37b8ca5a-e0f8-414b-a363-e56db520f027] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1083.222479] env[61728]: INFO nova.compute.manager [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Rebuilding instance [ 1083.269891] env[61728]: DEBUG nova.compute.manager [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1083.270792] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be91a51-74dd-4ebf-84c3-aba51862592e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.292852] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52807baa-c014-fd13-7dfb-a83088c68b10, 'name': SearchDatastore_Task, 'duration_secs': 0.014091} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.293642] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2cea929-a496-4a02-a438-6cceaef75e93 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.301176] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Waiting for the task: (returnval){ [ 1083.301176] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c2d739-ee66-580d-3abe-700256f206f5" [ 1083.301176] env[61728]: _type = "Task" [ 1083.301176] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.308761] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c2d739-ee66-580d-3abe-700256f206f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.426814] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526296c3-a74d-f9a7-c82e-4c9c4834fe20, 'name': SearchDatastore_Task, 'duration_secs': 0.016889} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.429154] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.429407] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1083.429622] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.451540] env[61728]: DEBUG oslo_vmware.api [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464845, 'name': PowerOffVM_Task, 'duration_secs': 0.281346} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.453879] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1083.454074] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1083.454873] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3637d800-d436-4d89-8696-d135afc59d2d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.458846] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.459089] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.459310] env[61728]: INFO nova.compute.manager [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Attaching volume 0136be7a-12a6-43a4-847f-203c70cf8dc8 to /dev/sdb [ 1083.464085] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6b41c4-c4b9-4712-b902-944dfaf5654c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.472633] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67eeb5fd-098c-4ac7-9a5c-3b15600ab4f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.504487] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f98be0-fc97-4683-924c-2476b6ba63f9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.507341] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd86ddc-1dd1-4de6-a15e-bb333e9fc96c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.516645] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bb88ca-4025-409d-a926-4801efb0c123 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.522038] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f469dd-a499-463b-a3d0-41453ccc7890 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.527592] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1083.528059] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1083.528406] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleting the datastore file [datastore1] 6c154517-68e8-4e70-86eb-cf53f6f812d8 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1083.529496] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-826c53d2-1299-4abc-8061-4c2e3f3e63e0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.540236] env[61728]: DEBUG nova.compute.provider_tree [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.548020] env[61728]: DEBUG nova.virt.block_device [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Updating existing volume attachment record: 2b83c425-0989-4e1d-9f4a-78a700d9281c {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1083.551556] env[61728]: DEBUG nova.scheduler.client.report [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1083.557101] env[61728]: DEBUG oslo_vmware.api [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for the task: (returnval){ [ 1083.557101] env[61728]: value = "task-464848" [ 1083.557101] env[61728]: _type = "Task" [ 1083.557101] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.565857] env[61728]: DEBUG oslo_vmware.api [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464848, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.609779] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 7b16fbc0-7f13-405f-b84e-e18de1ca7dd2] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1083.719615] env[61728]: DEBUG nova.network.neutron [req-fa4e724b-825f-4525-987b-c4035cb68c06 req-704375c3-4acc-406e-a2b6-766432af7023 service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Updated VIF entry in instance network info cache for port 4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1083.720065] env[61728]: DEBUG nova.network.neutron [req-fa4e724b-825f-4525-987b-c4035cb68c06 req-704375c3-4acc-406e-a2b6-766432af7023 service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Updating instance_info_cache with network_info: [{"id": "4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd", "address": "fa:16:3e:18:23:bd", "network": {"id": "22ad4b68-2895-4651-803f-44b8031a6292", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-489227259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "276942c8cf8a42729a541de096f69a0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c25f3d2-3e", "ovs_interfaceid": "4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.782616] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1083.783094] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-407ac06e-082b-4ffb-a8b3-ecda100d6ee1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.791515] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1083.791515] env[61728]: value = "task-464850" [ 1083.791515] env[61728]: _type = "Task" [ 1083.791515] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.800776] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464850, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.810576] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c2d739-ee66-580d-3abe-700256f206f5, 'name': SearchDatastore_Task, 'duration_secs': 0.011661} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.810842] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.811115] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 5eb3c3fa-d612-43d8-875c-2ee595e29081/5eb3c3fa-d612-43d8-875c-2ee595e29081.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1083.811392] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.811590] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1083.811829] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ef04673-4a97-4109-819d-b45ab8fce231 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.813931] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-17b4da84-a15b-4bc5-8f8b-54572609b8dd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.822459] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Waiting for the task: (returnval){ [ 1083.822459] env[61728]: value = "task-464851" [ 1083.822459] env[61728]: _type = "Task" [ 1083.822459] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.823767] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1083.823961] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1083.827460] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b411ff5-e166-456c-bb24-bf38f4271c13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.834670] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464851, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.835866] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1083.835866] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5219c863-1c5a-d159-a85f-e858657f1097" [ 1083.835866] env[61728]: _type = "Task" [ 1083.835866] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.843158] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5219c863-1c5a-d159-a85f-e858657f1097, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.891507] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1083.891807] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122206', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'name': 'volume-76406f6c-49da-4712-8faa-8af3a2f760a1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41329d0d-3838-4d1b-a4e6-8df4508eea76', 'attached_at': '', 'detached_at': '', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'serial': '76406f6c-49da-4712-8faa-8af3a2f760a1'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1083.892718] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b96ac40-b4a9-48c6-a898-f943215a1fe4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.908931] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08fa5b1-ebaf-4b73-8c12-2cbc1887ec7e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.932776] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] volume-76406f6c-49da-4712-8faa-8af3a2f760a1/volume-76406f6c-49da-4712-8faa-8af3a2f760a1.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1083.933122] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7166e320-43c4-45cd-a1f2-cb0b177171df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.951666] env[61728]: DEBUG oslo_vmware.api [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1083.951666] env[61728]: value = "task-464852" [ 1083.951666] env[61728]: _type = "Task" [ 1083.951666] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.961327] env[61728]: DEBUG oslo_vmware.api [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464852, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.058728] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.771s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.059343] env[61728]: DEBUG nova.compute.manager [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1084.072242] env[61728]: DEBUG oslo_vmware.api [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Task: {'id': task-464848, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321919} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.072556] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1084.072775] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1084.072982] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1084.073236] env[61728]: INFO nova.compute.manager [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1084.073512] env[61728]: DEBUG oslo.service.loopingcall [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1084.073740] env[61728]: DEBUG nova.compute.manager [-] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1084.073855] env[61728]: DEBUG nova.network.neutron [-] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1084.113868] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 4e5dc620-68a1-4e10-a8be-702c4999ca10] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1084.223345] env[61728]: DEBUG oslo_concurrency.lockutils [req-fa4e724b-825f-4525-987b-c4035cb68c06 req-704375c3-4acc-406e-a2b6-766432af7023 service nova] Releasing lock "refresh_cache-c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.306736] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464850, 'name': PowerOffVM_Task, 'duration_secs': 0.174094} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.307332] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1084.307740] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1084.309145] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b2cf34-c2ff-48ca-9c55-110b2c130786 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.318221] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1084.318531] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e0926ccc-a79e-4112-9714-bf179e0dbe26 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.332382] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464851, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.344785] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5219c863-1c5a-d159-a85f-e858657f1097, 'name': SearchDatastore_Task, 'duration_secs': 0.009946} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.345581] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74f48735-88ee-4e06-90cc-cafedeaa6c01 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.350628] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1084.350628] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f064e9-654d-c5bd-5188-7372103df697" [ 1084.350628] env[61728]: _type = "Task" [ 1084.350628] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.361045] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f064e9-654d-c5bd-5188-7372103df697, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.381806] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1084.382073] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1084.382270] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleting the datastore file [datastore1] 32a781b0-2db4-4883-ae96-d94c8181902b {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1084.382826] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1807ff4b-7e98-4b5e-9e5c-da4956f6e309 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.389778] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1084.389778] env[61728]: value = "task-464856" [ 1084.389778] env[61728]: _type = "Task" [ 1084.389778] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.398231] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464856, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.461977] env[61728]: DEBUG oslo_vmware.api [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.564741] env[61728]: DEBUG nova.compute.utils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1084.566225] env[61728]: DEBUG nova.compute.manager [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1084.566403] env[61728]: DEBUG nova.network.neutron [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1084.606123] env[61728]: DEBUG nova.policy [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54c30aeeea3a4c1b8ed34430361c2a4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a02bd852c564acd99e12fd17279f028', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1084.618480] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: a3e9af38-5115-4337-a20e-5d6c6ec655fe] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1084.834679] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464851, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516195} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.835174] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 5eb3c3fa-d612-43d8-875c-2ee595e29081/5eb3c3fa-d612-43d8-875c-2ee595e29081.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1084.835464] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1084.835782] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aae16e2a-7a4e-4147-8c21-34e90a80815b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.838357] env[61728]: DEBUG nova.network.neutron [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Successfully created port: ce80d456-9e6c-4d02-a7c6-99593a0761c6 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1084.845884] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Waiting for the task: (returnval){ [ 1084.845884] env[61728]: value = "task-464857" [ 1084.845884] env[61728]: _type = "Task" [ 1084.845884] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.853719] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464857, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.861851] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f064e9-654d-c5bd-5188-7372103df697, 'name': SearchDatastore_Task, 'duration_secs': 0.009827} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.862183] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.862503] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] c1ed0e8b-5cf9-4ff0-931e-fadc3432a125/c1ed0e8b-5cf9-4ff0-931e-fadc3432a125.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1084.862845] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c3589007-7768-49b1-9005-7084ec61f1ec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.870093] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1084.870093] env[61728]: value = "task-464858" [ 1084.870093] env[61728]: _type = "Task" [ 1084.870093] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.873065] env[61728]: DEBUG nova.network.neutron [-] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.879334] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464858, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.898102] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464856, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.244958} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.898425] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1084.898672] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1084.898915] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1084.963314] env[61728]: DEBUG oslo_vmware.api [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464852, 'name': ReconfigVM_Task, 'duration_secs': 0.609608} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.963685] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Reconfigured VM instance instance-0000006b to attach disk [datastore1] volume-76406f6c-49da-4712-8faa-8af3a2f760a1/volume-76406f6c-49da-4712-8faa-8af3a2f760a1.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.968530] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-26df8c1c-a72d-4257-9069-8492fad138c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.985294] env[61728]: DEBUG oslo_vmware.api [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1084.985294] env[61728]: value = "task-464859" [ 1084.985294] env[61728]: _type = "Task" [ 1084.985294] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.995133] env[61728]: DEBUG oslo_vmware.api [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464859, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.070508] env[61728]: DEBUG nova.compute.manager [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1085.121894] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 771341ed-8b8e-470c-9686-82650f5271b4] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1085.150034] env[61728]: DEBUG nova.compute.manager [req-0f52ef69-6135-4b42-9981-39287f320b81 req-3dfb8933-645d-41f7-ad20-4a5916e50c99 service nova] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Received event network-vif-deleted-e7f1c36b-89f0-4858-96a0-3c1875c4bb29 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1085.356464] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464857, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.376028] env[61728]: INFO nova.compute.manager [-] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Took 1.30 seconds to deallocate network for instance. [ 1085.380486] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464858, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.495562] env[61728]: DEBUG oslo_vmware.api [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464859, 'name': ReconfigVM_Task, 'duration_secs': 0.154677} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.495891] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122206', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'name': 'volume-76406f6c-49da-4712-8faa-8af3a2f760a1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41329d0d-3838-4d1b-a4e6-8df4508eea76', 'attached_at': '', 'detached_at': '', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'serial': '76406f6c-49da-4712-8faa-8af3a2f760a1'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1085.624699] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 9228fc5e-dbca-42b0-91cc-8e8d49f9eb8c] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1085.858570] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464857, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.880537] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464858, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.886041] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.886343] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.886545] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.914380] env[61728]: INFO nova.scheduler.client.report [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Deleted allocations for instance 6c154517-68e8-4e70-86eb-cf53f6f812d8 [ 1085.942979] env[61728]: DEBUG nova.virt.hardware [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1085.943301] env[61728]: DEBUG nova.virt.hardware [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1085.943472] env[61728]: DEBUG nova.virt.hardware [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1085.943678] env[61728]: DEBUG nova.virt.hardware [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1085.943833] env[61728]: DEBUG nova.virt.hardware [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1085.944101] env[61728]: DEBUG nova.virt.hardware [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1085.944225] env[61728]: DEBUG nova.virt.hardware [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1085.944389] env[61728]: DEBUG nova.virt.hardware [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1085.944557] env[61728]: DEBUG nova.virt.hardware [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1085.944737] env[61728]: DEBUG nova.virt.hardware [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1085.944891] env[61728]: DEBUG nova.virt.hardware [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1085.945864] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-625ec751-0a52-4fb1-bd5c-3226fc9d5e73 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.953997] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9e9fea-8783-42e1-bfa9-1b791830ab2d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.967843] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:84:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b80dd748-3d7e-4a23-a38d-9e79a3881452', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c4f8bc0c-ff22-428b-a962-d2d8b3672927', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1085.975250] env[61728]: DEBUG oslo.service.loopingcall [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1085.976224] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1085.976463] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b9dfe39-585c-4f14-8408-c09348b5213e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.995748] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1085.995748] env[61728]: value = "task-464860" [ 1085.995748] env[61728]: _type = "Task" [ 1085.995748] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.006170] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464860, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.080358] env[61728]: DEBUG nova.compute.manager [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1086.107142] env[61728]: DEBUG nova.virt.hardware [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1086.107448] env[61728]: DEBUG nova.virt.hardware [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1086.107609] env[61728]: DEBUG nova.virt.hardware [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1086.107800] env[61728]: DEBUG nova.virt.hardware [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1086.108038] env[61728]: DEBUG nova.virt.hardware [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1086.108119] env[61728]: DEBUG nova.virt.hardware [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1086.108326] env[61728]: DEBUG nova.virt.hardware [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1086.108551] env[61728]: DEBUG nova.virt.hardware [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1086.108712] env[61728]: DEBUG nova.virt.hardware [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1086.108888] env[61728]: DEBUG nova.virt.hardware [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1086.109081] env[61728]: DEBUG nova.virt.hardware [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1086.109977] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05644e8b-2e49-4223-b3fc-71552c1381fc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.119333] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed4a4d15-9816-488b-9261-c9f412ae2347 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.132792] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.133179] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Cleaning up deleted instances with incomplete migration {{(pid=61728) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1086.335575] env[61728]: DEBUG nova.network.neutron [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Successfully updated port: ce80d456-9e6c-4d02-a7c6-99593a0761c6 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1086.361494] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464857, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.124677} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.361804] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1086.363083] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1867d06e-d22c-4c36-92fd-bf1ed941e6ac {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.386725] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] 5eb3c3fa-d612-43d8-875c-2ee595e29081/5eb3c3fa-d612-43d8-875c-2ee595e29081.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.389747] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b6bde80-4ccd-470d-ba3c-73d12ee87978 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.408715] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464858, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.051644} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.409934] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] c1ed0e8b-5cf9-4ff0-931e-fadc3432a125/c1ed0e8b-5cf9-4ff0-931e-fadc3432a125.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1086.410128] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1086.410443] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Waiting for the task: (returnval){ [ 1086.410443] env[61728]: value = "task-464862" [ 1086.410443] env[61728]: _type = "Task" [ 1086.410443] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.410663] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-47c19108-f74b-4c27-8177-314fb72799c7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.426613] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1086.426613] env[61728]: value = "task-464863" [ 1086.426613] env[61728]: _type = "Task" [ 1086.426613] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.426613] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464862, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.426613] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cb94872-1394-4c25-b310-28f007380db9 tempest-DeleteServersTestJSON-2127553086 tempest-DeleteServersTestJSON-2127553086-project-member] Lock "6c154517-68e8-4e70-86eb-cf53f6f812d8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.507s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.437748] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464863, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.505952] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464860, 'name': CreateVM_Task, 'duration_secs': 0.318599} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.506158] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1086.506897] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.507082] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.507411] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1086.507670] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-714fcbc3-bf11-435e-a51a-3c81b10abbab {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.512579] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1086.512579] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522e9529-e347-826e-f249-ae06cbaa4b56" [ 1086.512579] env[61728]: _type = "Task" [ 1086.512579] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.520525] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522e9529-e347-826e-f249-ae06cbaa4b56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.535708] env[61728]: DEBUG nova.objects.instance [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lazy-loading 'flavor' on Instance uuid 41329d0d-3838-4d1b-a4e6-8df4508eea76 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.839148] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "refresh_cache-8ea37e93-16fc-4c60-9949-17656218d46a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.839148] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "refresh_cache-8ea37e93-16fc-4c60-9949-17656218d46a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.839148] env[61728]: DEBUG nova.network.neutron [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1086.922580] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464862, 'name': ReconfigVM_Task, 'duration_secs': 0.465872} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.922944] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Reconfigured VM instance instance-00000071 to attach disk [datastore1] 5eb3c3fa-d612-43d8-875c-2ee595e29081/5eb3c3fa-d612-43d8-875c-2ee595e29081.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.923594] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-254342b6-b2c0-433c-8417-f1c91b6d70a0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.931354] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Waiting for the task: (returnval){ [ 1086.931354] env[61728]: value = "task-464865" [ 1086.931354] env[61728]: _type = "Task" [ 1086.931354] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.937107] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464863, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062243} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.937619] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1086.938348] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d09d055-566b-40dc-a1d7-69f87d2f2520 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.943194] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464865, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.962399] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] c1ed0e8b-5cf9-4ff0-931e-fadc3432a125/c1ed0e8b-5cf9-4ff0-931e-fadc3432a125.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.962694] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c0bc3d7-0eeb-4f43-ac06-31f3592067ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.983528] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1086.983528] env[61728]: value = "task-464866" [ 1086.983528] env[61728]: _type = "Task" [ 1086.983528] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.992695] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464866, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.023480] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]522e9529-e347-826e-f249-ae06cbaa4b56, 'name': SearchDatastore_Task, 'duration_secs': 0.008169} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.023862] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.024151] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1087.024413] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.024588] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.024783] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1087.025104] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8a1243d4-1314-4dab-8e99-2447bb9cb77f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.033788] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1087.033934] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1087.034661] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e60daf1-a7d3-4259-a9c0-2da870e5db24 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.041971] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1087.041971] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c103bb-a5f1-2299-03bf-0f762184e32a" [ 1087.041971] env[61728]: _type = "Task" [ 1087.041971] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.042375] env[61728]: INFO nova.compute.manager [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Rebuilding instance [ 1087.047987] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7289fd01-de6b-4062-9cd6-f62dc9b48b79 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.762s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.053766] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c103bb-a5f1-2299-03bf-0f762184e32a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.092642] env[61728]: DEBUG nova.compute.manager [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1087.093626] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd91264-8d72-49f9-b5c9-23cafac37837 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.178405] env[61728]: DEBUG nova.compute.manager [req-0e3658ea-b7d0-47eb-9ae1-f4f5f2d367b3 req-36cd586d-5433-48ff-8568-146a6c45da56 service nova] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Received event network-vif-plugged-ce80d456-9e6c-4d02-a7c6-99593a0761c6 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1087.178666] env[61728]: DEBUG oslo_concurrency.lockutils [req-0e3658ea-b7d0-47eb-9ae1-f4f5f2d367b3 req-36cd586d-5433-48ff-8568-146a6c45da56 service nova] Acquiring lock "8ea37e93-16fc-4c60-9949-17656218d46a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.178879] env[61728]: DEBUG oslo_concurrency.lockutils [req-0e3658ea-b7d0-47eb-9ae1-f4f5f2d367b3 req-36cd586d-5433-48ff-8568-146a6c45da56 service nova] Lock "8ea37e93-16fc-4c60-9949-17656218d46a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.179043] env[61728]: DEBUG oslo_concurrency.lockutils [req-0e3658ea-b7d0-47eb-9ae1-f4f5f2d367b3 req-36cd586d-5433-48ff-8568-146a6c45da56 service nova] Lock "8ea37e93-16fc-4c60-9949-17656218d46a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.179197] env[61728]: DEBUG nova.compute.manager [req-0e3658ea-b7d0-47eb-9ae1-f4f5f2d367b3 req-36cd586d-5433-48ff-8568-146a6c45da56 service nova] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] No waiting events found dispatching network-vif-plugged-ce80d456-9e6c-4d02-a7c6-99593a0761c6 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1087.179371] env[61728]: WARNING nova.compute.manager [req-0e3658ea-b7d0-47eb-9ae1-f4f5f2d367b3 req-36cd586d-5433-48ff-8568-146a6c45da56 service nova] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Received unexpected event network-vif-plugged-ce80d456-9e6c-4d02-a7c6-99593a0761c6 for instance with vm_state building and task_state spawning. [ 1087.179534] env[61728]: DEBUG nova.compute.manager [req-0e3658ea-b7d0-47eb-9ae1-f4f5f2d367b3 req-36cd586d-5433-48ff-8568-146a6c45da56 service nova] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Received event network-changed-ce80d456-9e6c-4d02-a7c6-99593a0761c6 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1087.179705] env[61728]: DEBUG nova.compute.manager [req-0e3658ea-b7d0-47eb-9ae1-f4f5f2d367b3 req-36cd586d-5433-48ff-8568-146a6c45da56 service nova] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Refreshing instance network info cache due to event network-changed-ce80d456-9e6c-4d02-a7c6-99593a0761c6. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1087.179879] env[61728]: DEBUG oslo_concurrency.lockutils [req-0e3658ea-b7d0-47eb-9ae1-f4f5f2d367b3 req-36cd586d-5433-48ff-8568-146a6c45da56 service nova] Acquiring lock "refresh_cache-8ea37e93-16fc-4c60-9949-17656218d46a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.368162] env[61728]: DEBUG nova.network.neutron [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1087.441504] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464865, 'name': Rename_Task, 'duration_secs': 0.231695} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.441896] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1087.442184] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-264f2618-4387-43fc-8dfd-cb614c5a01e3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.448962] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Waiting for the task: (returnval){ [ 1087.448962] env[61728]: value = "task-464867" [ 1087.448962] env[61728]: _type = "Task" [ 1087.448962] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.456428] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464867, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.494251] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464866, 'name': ReconfigVM_Task, 'duration_secs': 0.285086} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.494584] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Reconfigured VM instance instance-00000072 to attach disk [datastore1] c1ed0e8b-5cf9-4ff0-931e-fadc3432a125/c1ed0e8b-5cf9-4ff0-931e-fadc3432a125.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.495179] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aca75cab-b41b-46ec-aa89-5c839a6068b9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.497388] env[61728]: DEBUG nova.network.neutron [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Updating instance_info_cache with network_info: [{"id": "ce80d456-9e6c-4d02-a7c6-99593a0761c6", "address": "fa:16:3e:32:ab:f0", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce80d456-9e", "ovs_interfaceid": "ce80d456-9e6c-4d02-a7c6-99593a0761c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.503352] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1087.503352] env[61728]: value = "task-464868" [ 1087.503352] env[61728]: _type = "Task" [ 1087.503352] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.511527] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464868, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.555135] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c103bb-a5f1-2299-03bf-0f762184e32a, 'name': SearchDatastore_Task, 'duration_secs': 0.014282} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.555935] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d121a97-3830-49c1-bee3-265d147cb337 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.561068] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1087.561068] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52252335-6799-a1be-b87b-33bd329f24e5" [ 1087.561068] env[61728]: _type = "Task" [ 1087.561068] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.568643] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52252335-6799-a1be-b87b-33bd329f24e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.605754] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1087.606313] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e901289c-1b70-42c9-892e-5aab2a9933cb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.613595] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1087.613595] env[61728]: value = "task-464869" [ 1087.613595] env[61728]: _type = "Task" [ 1087.613595] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.621656] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464869, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.958688] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464867, 'name': PowerOnVM_Task} progress is 76%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.000267] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "refresh_cache-8ea37e93-16fc-4c60-9949-17656218d46a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.000267] env[61728]: DEBUG nova.compute.manager [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Instance network_info: |[{"id": "ce80d456-9e6c-4d02-a7c6-99593a0761c6", "address": "fa:16:3e:32:ab:f0", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce80d456-9e", "ovs_interfaceid": "ce80d456-9e6c-4d02-a7c6-99593a0761c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1088.000610] env[61728]: DEBUG oslo_concurrency.lockutils [req-0e3658ea-b7d0-47eb-9ae1-f4f5f2d367b3 req-36cd586d-5433-48ff-8568-146a6c45da56 service nova] Acquired lock "refresh_cache-8ea37e93-16fc-4c60-9949-17656218d46a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.000852] env[61728]: DEBUG nova.network.neutron [req-0e3658ea-b7d0-47eb-9ae1-f4f5f2d367b3 req-36cd586d-5433-48ff-8568-146a6c45da56 service nova] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Refreshing network info cache for port ce80d456-9e6c-4d02-a7c6-99593a0761c6 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1088.002332] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:ab:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc6d5964-1106-4345-a26d-185dabd4ff0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ce80d456-9e6c-4d02-a7c6-99593a0761c6', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1088.011093] env[61728]: DEBUG oslo.service.loopingcall [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1088.012351] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1088.015675] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68669e10-3ebd-4927-9fa3-0b1b879e3ca1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.037407] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464868, 'name': Rename_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.038528] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1088.038528] env[61728]: value = "task-464870" [ 1088.038528] env[61728]: _type = "Task" [ 1088.038528] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.046128] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464870, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.071218] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52252335-6799-a1be-b87b-33bd329f24e5, 'name': SearchDatastore_Task, 'duration_secs': 0.020711} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.071470] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.071761] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 32a781b0-2db4-4883-ae96-d94c8181902b/32a781b0-2db4-4883-ae96-d94c8181902b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1088.072019] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc9cbb6a-6655-4da5-ae22-9e14e0af63d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.077886] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1088.077886] env[61728]: value = "task-464871" [ 1088.077886] env[61728]: _type = "Task" [ 1088.077886] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.085379] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464871, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.124658] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464869, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.136678] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "5fa2e9f4-6d68-46c2-a549-95d216bab886" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.136678] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "5fa2e9f4-6d68-46c2-a549-95d216bab886" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.136912] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "5fa2e9f4-6d68-46c2-a549-95d216bab886-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.137044] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "5fa2e9f4-6d68-46c2-a549-95d216bab886-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.137322] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "5fa2e9f4-6d68-46c2-a549-95d216bab886-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.139341] env[61728]: INFO nova.compute.manager [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Terminating instance [ 1088.141398] env[61728]: DEBUG nova.compute.manager [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1088.141645] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1088.142580] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbfcd0b-0310-40f1-b94f-50585e0c2aa0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.149614] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1088.149829] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33e09314-c340-433e-a3e7-56f143773d32 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.156650] env[61728]: DEBUG oslo_vmware.api [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1088.156650] env[61728]: value = "task-464872" [ 1088.156650] env[61728]: _type = "Task" [ 1088.156650] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.164146] env[61728]: DEBUG oslo_vmware.api [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464872, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.460767] env[61728]: DEBUG oslo_vmware.api [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464867, 'name': PowerOnVM_Task, 'duration_secs': 0.717313} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.461148] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1088.461316] env[61728]: INFO nova.compute.manager [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Took 10.59 seconds to spawn the instance on the hypervisor. [ 1088.461509] env[61728]: DEBUG nova.compute.manager [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1088.462523] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed721a03-947c-4d54-8ea6-7f87abc47ec2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.526405] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464868, 'name': Rename_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.548981] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464870, 'name': CreateVM_Task, 'duration_secs': 0.498478} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.551388] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1088.552125] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.552299] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.552621] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1088.553226] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae488b63-6583-4b49-be84-6216a961535a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.557831] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1088.557831] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5245aa42-691c-6b24-a227-ad8ceb770396" [ 1088.557831] env[61728]: _type = "Task" [ 1088.557831] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.566589] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5245aa42-691c-6b24-a227-ad8ceb770396, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.587608] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464871, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478171} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.587931] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 32a781b0-2db4-4883-ae96-d94c8181902b/32a781b0-2db4-4883-ae96-d94c8181902b.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1088.588175] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1088.588440] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-906f30fb-b46c-4411-adce-5fc54f89700f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.594421] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1088.594421] env[61728]: value = "task-464873" [ 1088.594421] env[61728]: _type = "Task" [ 1088.594421] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.602933] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464873, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.603896] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1088.604129] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122212', 'volume_id': '0136be7a-12a6-43a4-847f-203c70cf8dc8', 'name': 'volume-0136be7a-12a6-43a4-847f-203c70cf8dc8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd55beb02-00fc-4df6-a239-9e5d776a7c18', 'attached_at': '', 'detached_at': '', 'volume_id': '0136be7a-12a6-43a4-847f-203c70cf8dc8', 'serial': '0136be7a-12a6-43a4-847f-203c70cf8dc8'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1088.604910] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f36200-0415-4154-9078-95ef8fc744a4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.627498] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e12b6b3-c086-4bd7-98bc-6057937d9233 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.653731] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] volume-0136be7a-12a6-43a4-847f-203c70cf8dc8/volume-0136be7a-12a6-43a4-847f-203c70cf8dc8.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1088.658844] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c726cc0-a0ca-4664-9417-2e6a836329f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.672453] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464869, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.681688] env[61728]: DEBUG oslo_vmware.api [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464872, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.683150] env[61728]: DEBUG oslo_vmware.api [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1088.683150] env[61728]: value = "task-464874" [ 1088.683150] env[61728]: _type = "Task" [ 1088.683150] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.691681] env[61728]: DEBUG oslo_vmware.api [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464874, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.785326] env[61728]: DEBUG nova.network.neutron [req-0e3658ea-b7d0-47eb-9ae1-f4f5f2d367b3 req-36cd586d-5433-48ff-8568-146a6c45da56 service nova] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Updated VIF entry in instance network info cache for port ce80d456-9e6c-4d02-a7c6-99593a0761c6. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1088.785732] env[61728]: DEBUG nova.network.neutron [req-0e3658ea-b7d0-47eb-9ae1-f4f5f2d367b3 req-36cd586d-5433-48ff-8568-146a6c45da56 service nova] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Updating instance_info_cache with network_info: [{"id": "ce80d456-9e6c-4d02-a7c6-99593a0761c6", "address": "fa:16:3e:32:ab:f0", "network": {"id": "428a4270-763a-4765-b452-0f8931630e5c", "bridge": "br-int", "label": "tempest-ServersTestJSON-1068604966-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a02bd852c564acd99e12fd17279f028", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce80d456-9e", "ovs_interfaceid": "ce80d456-9e6c-4d02-a7c6-99593a0761c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.981027] env[61728]: INFO nova.compute.manager [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Took 18.18 seconds to build instance. [ 1089.023886] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464868, 'name': Rename_Task, 'duration_secs': 1.118362} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.024202] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1089.024469] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7be114a4-31e9-4ef5-9196-c777927e3709 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.031551] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1089.031551] env[61728]: value = "task-464875" [ 1089.031551] env[61728]: _type = "Task" [ 1089.031551] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.039520] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464875, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.068165] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5245aa42-691c-6b24-a227-ad8ceb770396, 'name': SearchDatastore_Task, 'duration_secs': 0.011353} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.068516] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.068825] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1089.069082] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.069243] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.069430] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1089.069695] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cf449a00-5f4d-46ba-b545-3889e8f21ccd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.085087] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1089.085281] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1089.085978] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d58a2b9d-087b-4f80-aade-e36e2ab22180 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.090952] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1089.090952] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5249ccc7-cc3d-8137-cc05-e80e0a57f202" [ 1089.090952] env[61728]: _type = "Task" [ 1089.090952] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.100531] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5249ccc7-cc3d-8137-cc05-e80e0a57f202, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.106398] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464873, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074086} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.106699] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1089.107455] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3f7f7c-8cf5-4254-b251-1d2d05affb3a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.130463] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] 32a781b0-2db4-4883-ae96-d94c8181902b/32a781b0-2db4-4883-ae96-d94c8181902b.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.133614] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1693517a-6fe2-46d4-973d-0b8d42218bbc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.153686] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464869, 'name': PowerOffVM_Task, 'duration_secs': 1.201934} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.154965] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1089.156636] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1089.156636] env[61728]: value = "task-464876" [ 1089.156636] env[61728]: _type = "Task" [ 1089.156636] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.165320] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464876, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.182096] env[61728]: DEBUG oslo_vmware.api [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464872, 'name': PowerOffVM_Task, 'duration_secs': 0.65905} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.182417] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1089.182600] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1089.182876] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d0f0c399-d19c-4c8a-ac60-56601c4046cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.193220] env[61728]: DEBUG oslo_vmware.api [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464874, 'name': ReconfigVM_Task, 'duration_secs': 0.364812} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.193501] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Reconfigured VM instance instance-00000069 to attach disk [datastore1] volume-0136be7a-12a6-43a4-847f-203c70cf8dc8/volume-0136be7a-12a6-43a4-847f-203c70cf8dc8.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1089.199344] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c872fc50-279a-4d03-8d1e-a64aa7312f7c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.212207] env[61728]: INFO nova.compute.manager [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Detaching volume 76406f6c-49da-4712-8faa-8af3a2f760a1 [ 1089.215653] env[61728]: DEBUG oslo_vmware.api [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1089.215653] env[61728]: value = "task-464878" [ 1089.215653] env[61728]: _type = "Task" [ 1089.215653] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.226531] env[61728]: DEBUG oslo_vmware.api [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464878, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.249247] env[61728]: INFO nova.virt.block_device [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Attempting to driver detach volume 76406f6c-49da-4712-8faa-8af3a2f760a1 from mountpoint /dev/sdb [ 1089.249569] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1089.249698] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122206', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'name': 'volume-76406f6c-49da-4712-8faa-8af3a2f760a1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41329d0d-3838-4d1b-a4e6-8df4508eea76', 'attached_at': '', 'detached_at': '', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'serial': '76406f6c-49da-4712-8faa-8af3a2f760a1'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1089.250765] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d21f50d3-e396-4619-9bfd-b2cc7d17d73b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.254534] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1089.254750] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1089.254986] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Deleting the datastore file [datastore1] 5fa2e9f4-6d68-46c2-a549-95d216bab886 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1089.255688] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8515edb5-c03c-45c7-b9e4-522ddc313d66 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.286313] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9289829f-80b1-4242-89f5-94d9bd2dacd7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.290573] env[61728]: DEBUG oslo_vmware.api [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1089.290573] env[61728]: value = "task-464879" [ 1089.290573] env[61728]: _type = "Task" [ 1089.290573] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.291214] env[61728]: DEBUG oslo_concurrency.lockutils [req-0e3658ea-b7d0-47eb-9ae1-f4f5f2d367b3 req-36cd586d-5433-48ff-8568-146a6c45da56 service nova] Releasing lock "refresh_cache-8ea37e93-16fc-4c60-9949-17656218d46a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.299413] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a474175e-711e-4232-9710-14f8b10bfcfa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.307344] env[61728]: DEBUG oslo_vmware.api [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464879, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.326096] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea5dfd9-e9fb-433d-87ef-b58738a41774 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.342581] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] The volume has not been displaced from its original location: [datastore1] volume-76406f6c-49da-4712-8faa-8af3a2f760a1/volume-76406f6c-49da-4712-8faa-8af3a2f760a1.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1089.347934] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1089.348383] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.348610] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.348783] env[61728]: INFO nova.compute.manager [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Shelving [ 1089.350150] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90913874-247f-44f1-99e4-6ee0200cd8f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.376265] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1089.376265] env[61728]: value = "task-464880" [ 1089.376265] env[61728]: _type = "Task" [ 1089.376265] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.384814] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.482760] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3308e5aa-8575-4f2f-a111-c81f5ee0171e tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Lock "5eb3c3fa-d612-43d8-875c-2ee595e29081" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.699s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.541761] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464875, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.602468] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5249ccc7-cc3d-8137-cc05-e80e0a57f202, 'name': SearchDatastore_Task, 'duration_secs': 0.019875} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.603268] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13b39d3b-bfd8-4e84-b86e-57b1b83934bf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.609458] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1089.609458] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52076eca-04b7-ce23-c63c-621943856011" [ 1089.609458] env[61728]: _type = "Task" [ 1089.609458] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.618363] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52076eca-04b7-ce23-c63c-621943856011, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.667602] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464876, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.726160] env[61728]: DEBUG oslo_vmware.api [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464878, 'name': ReconfigVM_Task, 'duration_secs': 0.152033} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.726491] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122212', 'volume_id': '0136be7a-12a6-43a4-847f-203c70cf8dc8', 'name': 'volume-0136be7a-12a6-43a4-847f-203c70cf8dc8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd55beb02-00fc-4df6-a239-9e5d776a7c18', 'attached_at': '', 'detached_at': '', 'volume_id': '0136be7a-12a6-43a4-847f-203c70cf8dc8', 'serial': '0136be7a-12a6-43a4-847f-203c70cf8dc8'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1089.802427] env[61728]: DEBUG oslo_vmware.api [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464879, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.486378} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.802698] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1089.802922] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1089.803129] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1089.803316] env[61728]: INFO nova.compute.manager [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Took 1.66 seconds to destroy the instance on the hypervisor. [ 1089.803571] env[61728]: DEBUG oslo.service.loopingcall [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1089.803774] env[61728]: DEBUG nova.compute.manager [-] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1089.803871] env[61728]: DEBUG nova.network.neutron [-] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1089.872171] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1089.872453] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1c2c51d-36d2-4a4f-bf11-80325ddd164d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.881144] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1089.881144] env[61728]: value = "task-464881" [ 1089.881144] env[61728]: _type = "Task" [ 1089.881144] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.887618] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464880, 'name': ReconfigVM_Task, 'duration_secs': 0.472769} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.888318] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1089.896811] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1db7181b-d18b-4ffe-bb56-4bbacc301114 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.906187] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464881, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.912140] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1089.912140] env[61728]: value = "task-464882" [ 1089.912140] env[61728]: _type = "Task" [ 1089.912140] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.921686] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464882, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.929296] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Acquiring lock "5eb3c3fa-d612-43d8-875c-2ee595e29081" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.929553] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Lock "5eb3c3fa-d612-43d8-875c-2ee595e29081" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.929771] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Acquiring lock "5eb3c3fa-d612-43d8-875c-2ee595e29081-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.929959] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Lock "5eb3c3fa-d612-43d8-875c-2ee595e29081-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.930153] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Lock "5eb3c3fa-d612-43d8-875c-2ee595e29081-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.933056] env[61728]: INFO nova.compute.manager [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Terminating instance [ 1089.935466] env[61728]: DEBUG nova.compute.manager [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1089.935761] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1089.936552] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ee3536-b576-4650-bdca-916a772ef9a0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.944496] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1089.944748] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b6cb4467-18c6-4670-add4-01d35b791369 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.952576] env[61728]: DEBUG oslo_vmware.api [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Waiting for the task: (returnval){ [ 1089.952576] env[61728]: value = "task-464883" [ 1089.952576] env[61728]: _type = "Task" [ 1089.952576] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.961031] env[61728]: DEBUG oslo_vmware.api [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464883, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.044475] env[61728]: DEBUG oslo_vmware.api [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464875, 'name': PowerOnVM_Task, 'duration_secs': 0.798766} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.044730] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1090.044894] env[61728]: INFO nova.compute.manager [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Took 9.85 seconds to spawn the instance on the hypervisor. [ 1090.045126] env[61728]: DEBUG nova.compute.manager [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1090.045981] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f11544-a468-46b4-9104-33e1363b7124 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.123966] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1090.126212] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52076eca-04b7-ce23-c63c-621943856011, 'name': SearchDatastore_Task, 'duration_secs': 0.019313} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.127391] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.127839] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 8ea37e93-16fc-4c60-9949-17656218d46a/8ea37e93-16fc-4c60-9949-17656218d46a.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1090.128262] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e6353ba-78de-4675-b25b-fa4ab5f27ca4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.142823] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1090.142823] env[61728]: value = "task-464884" [ 1090.142823] env[61728]: _type = "Task" [ 1090.142823] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.151980] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464884, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.168508] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464876, 'name': ReconfigVM_Task, 'duration_secs': 0.633161} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.169733] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Reconfigured VM instance instance-00000070 to attach disk [datastore1] 32a781b0-2db4-4883-ae96-d94c8181902b/32a781b0-2db4-4883-ae96-d94c8181902b.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1090.170418] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c0cb5f7e-0ced-4601-b0dd-81fc67153bb3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.176750] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1090.176750] env[61728]: value = "task-464885" [ 1090.176750] env[61728]: _type = "Task" [ 1090.176750] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.186349] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464885, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.245066] env[61728]: DEBUG nova.compute.manager [req-eba9ae62-646d-4378-8489-37fb12475c19 req-958a61c2-7c1c-4be5-b1fd-6e204f445db8 service nova] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Received event network-vif-deleted-32361465-1ad4-4a38-9989-8110f57692f7 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1090.245066] env[61728]: INFO nova.compute.manager [req-eba9ae62-646d-4378-8489-37fb12475c19 req-958a61c2-7c1c-4be5-b1fd-6e204f445db8 service nova] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Neutron deleted interface 32361465-1ad4-4a38-9989-8110f57692f7; detaching it from the instance and deleting it from the info cache [ 1090.245222] env[61728]: DEBUG nova.network.neutron [req-eba9ae62-646d-4378-8489-37fb12475c19 req-958a61c2-7c1c-4be5-b1fd-6e204f445db8 service nova] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.392842] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464881, 'name': PowerOffVM_Task, 'duration_secs': 0.216754} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.393185] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1090.394121] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b885ccb-4bae-4009-b755-3a182a4b5229 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.418053] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98def2b3-b8b1-467e-b987-2912846ece65 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.428721] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464882, 'name': ReconfigVM_Task, 'duration_secs': 0.265203} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.430969] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122206', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'name': 'volume-76406f6c-49da-4712-8faa-8af3a2f760a1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41329d0d-3838-4d1b-a4e6-8df4508eea76', 'attached_at': '', 'detached_at': '', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'serial': '76406f6c-49da-4712-8faa-8af3a2f760a1'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1090.465655] env[61728]: DEBUG oslo_vmware.api [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464883, 'name': PowerOffVM_Task, 'duration_secs': 0.18215} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.465989] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1090.466154] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1090.466648] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-64148772-2ddf-443a-8acf-f08f4c4cc67c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.556836] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1090.557125] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1090.557170] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Deleting the datastore file [datastore1] 5eb3c3fa-d612-43d8-875c-2ee595e29081 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1090.557766] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78ea38b8-c21b-43d5-9fec-449c6c4ca1d4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.569111] env[61728]: INFO nova.compute.manager [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Took 18.50 seconds to build instance. [ 1090.570990] env[61728]: DEBUG oslo_vmware.api [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Waiting for the task: (returnval){ [ 1090.570990] env[61728]: value = "task-464887" [ 1090.570990] env[61728]: _type = "Task" [ 1090.570990] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.586758] env[61728]: DEBUG oslo_vmware.api [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464887, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.630852] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1090.656414] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464884, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.688749] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464885, 'name': Rename_Task, 'duration_secs': 0.211745} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.689144] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1090.689420] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e06d9df1-5745-40c7-ae0d-4eb06d12c935 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.697520] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1090.697520] env[61728]: value = "task-464888" [ 1090.697520] env[61728]: _type = "Task" [ 1090.697520] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.706644] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464888, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.712107] env[61728]: DEBUG nova.network.neutron [-] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.749498] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5f9f5e1-0e6d-466c-8d2e-bbef38ea829d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.759679] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de0b7ed-29f5-4110-b4ba-027c425c9b13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.773997] env[61728]: DEBUG nova.objects.instance [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lazy-loading 'flavor' on Instance uuid d55beb02-00fc-4df6-a239-9e5d776a7c18 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1090.797523] env[61728]: DEBUG nova.compute.manager [req-eba9ae62-646d-4378-8489-37fb12475c19 req-958a61c2-7c1c-4be5-b1fd-6e204f445db8 service nova] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Detach interface failed, port_id=32361465-1ad4-4a38-9989-8110f57692f7, reason: Instance 5fa2e9f4-6d68-46c2-a549-95d216bab886 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1090.938451] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1090.938801] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bf4380a3-a913-4d9f-8224-a4617334198a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.947354] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1090.947354] env[61728]: value = "task-464889" [ 1090.947354] env[61728]: _type = "Task" [ 1090.947354] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.956669] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464889, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.076521] env[61728]: DEBUG oslo_concurrency.lockutils [None req-00a86eb2-bdc4-4fa5-9d6d-60c9dd65e479 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.020s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.091127] env[61728]: DEBUG oslo_vmware.api [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Task: {'id': task-464887, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.444627} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.091437] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1091.091689] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1091.091952] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1091.092199] env[61728]: INFO nova.compute.manager [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1091.092493] env[61728]: DEBUG oslo.service.loopingcall [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.092734] env[61728]: DEBUG nova.compute.manager [-] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1091.092867] env[61728]: DEBUG nova.network.neutron [-] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1091.139747] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Getting list of instances from cluster (obj){ [ 1091.139747] env[61728]: value = "domain-c8" [ 1091.139747] env[61728]: _type = "ClusterComputeResource" [ 1091.139747] env[61728]: } {{(pid=61728) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1091.143784] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09aa0c82-bd49-474c-90c6-073b51113d3f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.157517] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464884, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.172031] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 8ea37e93-16fc-4c60-9949-17656218d46a/8ea37e93-16fc-4c60-9949-17656218d46a.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1091.172031] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1091.172031] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Got total of 11 instances {{(pid=61728) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1091.172273] env[61728]: WARNING nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] While synchronizing instance power states, found 13 instances in the database and 11 instances on the hypervisor. [ 1091.172434] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Triggering sync for uuid 40ef29df-12d6-4274-846a-4a2c32752b8b {{(pid=61728) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1091.172582] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Triggering sync for uuid f27f22aa-f05d-4981-a389-311731c51f93 {{(pid=61728) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1091.172830] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Triggering sync for uuid 5fa2e9f4-6d68-46c2-a549-95d216bab886 {{(pid=61728) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1091.174887] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Triggering sync for uuid b1fb2e3f-234a-44c7-983e-d4441b4e3cef {{(pid=61728) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1091.174887] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Triggering sync for uuid 8c5b2380-bd14-451e-9612-2767c2a6adb2 {{(pid=61728) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1091.174887] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Triggering sync for uuid d55beb02-00fc-4df6-a239-9e5d776a7c18 {{(pid=61728) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1091.174887] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Triggering sync for uuid 41329d0d-3838-4d1b-a4e6-8df4508eea76 {{(pid=61728) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1091.174887] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Triggering sync for uuid 77b8f509-17e7-4c25-a43d-1655c16ce8d1 {{(pid=61728) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1091.174887] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Triggering sync for uuid 532507fa-996d-45c7-bf70-f09c93be79ed {{(pid=61728) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1091.174887] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Triggering sync for uuid 32a781b0-2db4-4883-ae96-d94c8181902b {{(pid=61728) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1091.174887] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Triggering sync for uuid 5eb3c3fa-d612-43d8-875c-2ee595e29081 {{(pid=61728) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1091.174887] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Triggering sync for uuid c1ed0e8b-5cf9-4ff0-931e-fadc3432a125 {{(pid=61728) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1091.174887] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Triggering sync for uuid 8ea37e93-16fc-4c60-9949-17656218d46a {{(pid=61728) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1091.174887] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e9dde8a2-459e-4337-a68e-73d0096b2337 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.180732] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "40ef29df-12d6-4274-846a-4a2c32752b8b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.180974] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "40ef29df-12d6-4274-846a-4a2c32752b8b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.181267] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "f27f22aa-f05d-4981-a389-311731c51f93" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.181455] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "f27f22aa-f05d-4981-a389-311731c51f93" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.181686] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "5fa2e9f4-6d68-46c2-a549-95d216bab886" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.182285] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.182557] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.182962] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.183581] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.185252] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.185746] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.185996] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.186230] env[61728]: INFO nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] During sync_power_state the instance has a pending task (rebuilding). Skip. [ 1091.186440] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.186692] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.186950] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "532507fa-996d-45c7-bf70-f09c93be79ed" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.187189] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "532507fa-996d-45c7-bf70-f09c93be79ed" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.187470] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "32a781b0-2db4-4883-ae96-d94c8181902b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.187699] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "32a781b0-2db4-4883-ae96-d94c8181902b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.188113] env[61728]: INFO nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] During sync_power_state the instance has a pending task (rebuild_spawning). Skip. [ 1091.188113] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "32a781b0-2db4-4883-ae96-d94c8181902b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.188338] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "5eb3c3fa-d612-43d8-875c-2ee595e29081" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.188593] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.188812] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.189098] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "8ea37e93-16fc-4c60-9949-17656218d46a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.190184] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b99b7c-1219-469b-94d7-1844e916b0be {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.195542] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b28bab43-a55b-45f5-a047-473fc160e1f9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.196471] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f1316b-7613-40b7-9500-cb52f000a37e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.199770] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0255e8b7-635a-4a0b-9518-13d9365a23f7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.202105] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20299b9-26bd-4477-9ea3-101bb70dc1b9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.205634] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381a4a43-3b9d-4401-9a3f-732acf780f66 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.219650] env[61728]: INFO nova.compute.manager [-] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Took 1.42 seconds to deallocate network for instance. [ 1091.236949] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1091.236949] env[61728]: value = "task-464890" [ 1091.236949] env[61728]: _type = "Task" [ 1091.236949] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.266687] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464888, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.272123] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464890, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.282787] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ed67ce83-58cb-4ca7-b401-efa1f41f418b tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.823s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.283410] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.098s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.285424] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c316eab0-5806-4b4f-ba00-c473dc4527ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.457377] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464889, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.495264] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1091.495264] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d6d6f9fa-0f7f-4a28-aa6b-e4c06522c6de {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.502337] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1091.502337] env[61728]: value = "task-464891" [ 1091.502337] env[61728]: _type = "Task" [ 1091.502337] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.514616] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1091.514616] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1091.514616] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122206', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'name': 'volume-76406f6c-49da-4712-8faa-8af3a2f760a1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41329d0d-3838-4d1b-a4e6-8df4508eea76', 'attached_at': '', 'detached_at': '', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'serial': '76406f6c-49da-4712-8faa-8af3a2f760a1'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1091.514616] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6299f80-6c61-492d-aa72-634ed6547d93 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.533993] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b6fce8-14b6-44a5-ab9a-6edbccf1831c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.542041] env[61728]: WARNING nova.virt.vmwareapi.driver [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1091.542396] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1091.543261] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02cabf74-6de2-45a2-9d1e-0a1f999c28a8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.551175] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1091.551437] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-83ac533c-4071-4fe6-b43e-0ecc7c0d7e75 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.619518] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1091.619852] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1091.620154] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleting the datastore file [datastore1] 41329d0d-3838-4d1b-a4e6-8df4508eea76 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1091.620435] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad7dc827-864f-4117-875f-4eefb6407e70 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.627759] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1091.627759] env[61728]: value = "task-464893" [ 1091.627759] env[61728]: _type = "Task" [ 1091.627759] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.635377] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464893, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.723363] env[61728]: DEBUG oslo_vmware.api [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464888, 'name': PowerOnVM_Task, 'duration_secs': 0.813017} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.723645] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1091.724554] env[61728]: DEBUG nova.compute.manager [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1091.724637] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4eb4b7b-bb72-448f-b532-44eae3d3cb16 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.746627] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "f27f22aa-f05d-4981-a389-311731c51f93" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.565s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.752548] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.752818] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.753056] env[61728]: DEBUG nova.objects.instance [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lazy-loading 'resources' on Instance uuid 5fa2e9f4-6d68-46c2-a549-95d216bab886 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.754616] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "40ef29df-12d6-4274-846a-4a2c32752b8b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.574s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.755133] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.571s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.761890] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.579s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.762014] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464890, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069075} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.762851] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.574s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.763107] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1091.763913] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb36e78-6ae2-446d-9488-a86bf7428e8e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.766879] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "532507fa-996d-45c7-bf70-f09c93be79ed" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.580s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.787135] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] 8ea37e93-16fc-4c60-9949-17656218d46a/8ea37e93-16fc-4c60-9949-17656218d46a.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1091.788288] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8aa5558a-1ad3-4bfc-a482-cc7507de7521 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.804762] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.521s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.811447] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1091.811447] env[61728]: value = "task-464894" [ 1091.811447] env[61728]: _type = "Task" [ 1091.811447] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.819810] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464894, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.902134] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd2b2e21-84ab-4a9f-82ce-15901e857bc8 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.902440] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd2b2e21-84ab-4a9f-82ce-15901e857bc8 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.902638] env[61728]: DEBUG nova.compute.manager [None req-cd2b2e21-84ab-4a9f-82ce-15901e857bc8 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1091.908058] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d3eeb0-e3d9-452e-bdb3-d71b5837b265 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.914986] env[61728]: DEBUG nova.compute.manager [None req-cd2b2e21-84ab-4a9f-82ce-15901e857bc8 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61728) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1091.915759] env[61728]: DEBUG nova.objects.instance [None req-cd2b2e21-84ab-4a9f-82ce-15901e857bc8 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lazy-loading 'flavor' on Instance uuid d55beb02-00fc-4df6-a239-9e5d776a7c18 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.960106] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464889, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.037678] env[61728]: DEBUG nova.network.neutron [-] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.138637] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464893, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.430451} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.138897] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1092.139125] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1092.139311] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1092.243019] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.291996] env[61728]: DEBUG nova.compute.manager [req-9828f6c6-32ad-43a2-81c7-761e60ea79d9 req-e11933ce-6d48-4710-bb5f-00445675b1b0 service nova] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Received event network-vif-deleted-e8e000ef-5e57-43c4-9e1c-6d296929f225 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1092.292551] env[61728]: DEBUG nova.compute.manager [req-9828f6c6-32ad-43a2-81c7-761e60ea79d9 req-e11933ce-6d48-4710-bb5f-00445675b1b0 service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Received event network-changed-4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1092.292551] env[61728]: DEBUG nova.compute.manager [req-9828f6c6-32ad-43a2-81c7-761e60ea79d9 req-e11933ce-6d48-4710-bb5f-00445675b1b0 service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Refreshing instance network info cache due to event network-changed-4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1092.292692] env[61728]: DEBUG oslo_concurrency.lockutils [req-9828f6c6-32ad-43a2-81c7-761e60ea79d9 req-e11933ce-6d48-4710-bb5f-00445675b1b0 service nova] Acquiring lock "refresh_cache-c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.292914] env[61728]: DEBUG oslo_concurrency.lockutils [req-9828f6c6-32ad-43a2-81c7-761e60ea79d9 req-e11933ce-6d48-4710-bb5f-00445675b1b0 service nova] Acquired lock "refresh_cache-c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.293147] env[61728]: DEBUG nova.network.neutron [req-9828f6c6-32ad-43a2-81c7-761e60ea79d9 req-e11933ce-6d48-4710-bb5f-00445675b1b0 service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Refreshing network info cache for port 4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1092.323255] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464894, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.420833] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd2b2e21-84ab-4a9f-82ce-15901e857bc8 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1092.421170] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c496a7b6-d236-4783-9a4b-aaa8f31fe21b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.431661] env[61728]: DEBUG oslo_vmware.api [None req-cd2b2e21-84ab-4a9f-82ce-15901e857bc8 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1092.431661] env[61728]: value = "task-464895" [ 1092.431661] env[61728]: _type = "Task" [ 1092.431661] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.442441] env[61728]: DEBUG oslo_vmware.api [None req-cd2b2e21-84ab-4a9f-82ce-15901e857bc8 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464895, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.460044] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464889, 'name': CreateSnapshot_Task, 'duration_secs': 1.109001} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.462941] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1092.464054] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3ccb01-1cd9-459a-9cef-d6cd11809902 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.481326] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77c48fb-1197-401f-a0f3-f90ee68bbc3e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.491143] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d02b8e-b08c-49b2-83b1-e363efdc7cde {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.528258] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe22a12-233c-4c2f-986c-3b3443a578c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.538542] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-619839dc-3ea5-426a-8511-5f6658ed3e27 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.543359] env[61728]: INFO nova.compute.manager [-] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Took 1.45 seconds to deallocate network for instance. [ 1092.559387] env[61728]: DEBUG nova.compute.provider_tree [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1092.644983] env[61728]: INFO nova.virt.block_device [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Booting with volume 76406f6c-49da-4712-8faa-8af3a2f760a1 at /dev/sdb [ 1092.686809] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-821060cf-78d1-475d-81a7-580586d26bfd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.697521] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd059bad-51d3-4554-a186-8d977c0ced5d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.730599] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5ce8f26c-0ef5-43a7-a0d0-dd2b995ebdd0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.739777] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f9d0ff-abfe-4d4e-b093-625112da9646 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.776910] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e499cd38-ea8f-492e-9537-c531a5adaa08 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.784641] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a9596b-6d7c-494f-a799-3b139c787359 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.801511] env[61728]: DEBUG nova.virt.block_device [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Updating existing volume attachment record: f3a466d5-7e2e-4552-8d87-620dad31cf0e {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1092.822924] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464894, 'name': ReconfigVM_Task, 'duration_secs': 0.57791} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.824312] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Reconfigured VM instance instance-00000073 to attach disk [datastore1] 8ea37e93-16fc-4c60-9949-17656218d46a/8ea37e93-16fc-4c60-9949-17656218d46a.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1092.824993] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-73fa10ff-bade-4695-bc58-bd4604065806 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.832761] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1092.832761] env[61728]: value = "task-464896" [ 1092.832761] env[61728]: _type = "Task" [ 1092.832761] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.843176] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464896, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.941883] env[61728]: DEBUG oslo_vmware.api [None req-cd2b2e21-84ab-4a9f-82ce-15901e857bc8 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464895, 'name': PowerOffVM_Task, 'duration_secs': 0.197205} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.942196] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd2b2e21-84ab-4a9f-82ce-15901e857bc8 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1092.942385] env[61728]: DEBUG nova.compute.manager [None req-cd2b2e21-84ab-4a9f-82ce-15901e857bc8 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1092.943196] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8899f11b-b650-4f9a-986d-b1f4cc2ac089 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.993351] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1092.993778] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7e863cc7-472b-4f90-a3c8-69dc529d06e5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.009024] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1093.009024] env[61728]: value = "task-464897" [ 1093.009024] env[61728]: _type = "Task" [ 1093.009024] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.019355] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464897, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.060443] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.061371] env[61728]: DEBUG nova.scheduler.client.report [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1093.162597] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "32a781b0-2db4-4883-ae96-d94c8181902b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.163222] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "32a781b0-2db4-4883-ae96-d94c8181902b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.163222] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "32a781b0-2db4-4883-ae96-d94c8181902b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.163408] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "32a781b0-2db4-4883-ae96-d94c8181902b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.163596] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "32a781b0-2db4-4883-ae96-d94c8181902b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.165901] env[61728]: INFO nova.compute.manager [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Terminating instance [ 1093.167954] env[61728]: DEBUG nova.compute.manager [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1093.168180] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1093.169028] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0e71c2-395e-4b3b-8ef6-8f6c0a86ad04 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.174923] env[61728]: DEBUG nova.network.neutron [req-9828f6c6-32ad-43a2-81c7-761e60ea79d9 req-e11933ce-6d48-4710-bb5f-00445675b1b0 service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Updated VIF entry in instance network info cache for port 4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1093.175298] env[61728]: DEBUG nova.network.neutron [req-9828f6c6-32ad-43a2-81c7-761e60ea79d9 req-e11933ce-6d48-4710-bb5f-00445675b1b0 service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Updating instance_info_cache with network_info: [{"id": "4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd", "address": "fa:16:3e:18:23:bd", "network": {"id": "22ad4b68-2895-4651-803f-44b8031a6292", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-489227259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.236", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "276942c8cf8a42729a541de096f69a0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c25f3d2-3e", "ovs_interfaceid": "4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.178880] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1093.179418] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c2193e7f-ef20-47e7-8593-1cfa53920791 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.187994] env[61728]: DEBUG oslo_vmware.api [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1093.187994] env[61728]: value = "task-464898" [ 1093.187994] env[61728]: _type = "Task" [ 1093.187994] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.197528] env[61728]: DEBUG oslo_vmware.api [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464898, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.348326] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464896, 'name': Rename_Task, 'duration_secs': 0.479026} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.348326] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1093.348326] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c1189212-a0e8-45ad-9818-07ef9f95341b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.358263] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1093.358263] env[61728]: value = "task-464899" [ 1093.358263] env[61728]: _type = "Task" [ 1093.358263] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.368786] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464899, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.457078] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cd2b2e21-84ab-4a9f-82ce-15901e857bc8 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.554s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.524886] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464897, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.566675] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.814s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.569920] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.326s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.569920] env[61728]: DEBUG nova.objects.instance [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61728) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1093.597114] env[61728]: INFO nova.scheduler.client.report [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Deleted allocations for instance 5fa2e9f4-6d68-46c2-a549-95d216bab886 [ 1093.680403] env[61728]: DEBUG oslo_concurrency.lockutils [req-9828f6c6-32ad-43a2-81c7-761e60ea79d9 req-e11933ce-6d48-4710-bb5f-00445675b1b0 service nova] Releasing lock "refresh_cache-c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.700941] env[61728]: DEBUG oslo_vmware.api [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464898, 'name': PowerOffVM_Task, 'duration_secs': 0.291281} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.701422] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1093.701691] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1093.702012] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c93702fb-da5c-413e-b68e-f33603dde665 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.772764] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1093.773113] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1093.773378] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleting the datastore file [datastore1] 32a781b0-2db4-4883-ae96-d94c8181902b {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1093.774229] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ee74c8c-2681-4ec0-b546-dff15b9144d8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.783607] env[61728]: DEBUG oslo_vmware.api [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1093.783607] env[61728]: value = "task-464901" [ 1093.783607] env[61728]: _type = "Task" [ 1093.783607] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.801019] env[61728]: DEBUG oslo_vmware.api [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464901, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.873948] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464899, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.021948] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464897, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.110552] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bff972a5-937a-4568-9a56-81e0d5a8d8e5 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "5fa2e9f4-6d68-46c2-a549-95d216bab886" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.974s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.111540] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "5fa2e9f4-6d68-46c2-a549-95d216bab886" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.930s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.111704] env[61728]: INFO nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] During sync_power_state the instance has a pending task (deleting). Skip. [ 1094.111937] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "5fa2e9f4-6d68-46c2-a549-95d216bab886" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.294962] env[61728]: DEBUG oslo_vmware.api [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464901, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164369} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.295274] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1094.295469] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1094.295654] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1094.295832] env[61728]: INFO nova.compute.manager [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1094.296098] env[61728]: DEBUG oslo.service.loopingcall [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1094.296304] env[61728]: DEBUG nova.compute.manager [-] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1094.296403] env[61728]: DEBUG nova.network.neutron [-] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1094.372150] env[61728]: DEBUG oslo_vmware.api [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464899, 'name': PowerOnVM_Task, 'duration_secs': 0.589636} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.372590] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1094.372910] env[61728]: INFO nova.compute.manager [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Took 8.29 seconds to spawn the instance on the hypervisor. [ 1094.373185] env[61728]: DEBUG nova.compute.manager [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1094.373979] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e24de4b-9a71-4c42-afe4-50cc52060e15 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.437627] env[61728]: DEBUG nova.objects.instance [None req-bdbb147a-5737-413d-81de-0827f217e7fd tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lazy-loading 'flavor' on Instance uuid d55beb02-00fc-4df6-a239-9e5d776a7c18 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.524278] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464897, 'name': CloneVM_Task, 'duration_secs': 1.383884} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.524663] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Created linked-clone VM from snapshot [ 1094.525326] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce5203f9-5f4e-47dd-86cf-b719c53bd26d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.535499] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Uploading image d9914484-11b1-4819-982a-67a108bd3ed1 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1094.576089] env[61728]: DEBUG oslo_vmware.rw_handles [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1094.576089] env[61728]: value = "vm-122216" [ 1094.576089] env[61728]: _type = "VirtualMachine" [ 1094.576089] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1094.577649] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-cd77c55f-6493-4de6-86dc-a59e8d43a592 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.587140] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d1f4222b-d925-44d2-a2b0-fb98c3ae4c55 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.588995] env[61728]: DEBUG nova.compute.manager [req-c3c814e9-7de1-48e4-8eda-25b4997a1075 req-a0b4fffd-f2cb-49c0-ab00-1dfeb2a88443 service nova] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Received event network-vif-deleted-c4f8bc0c-ff22-428b-a962-d2d8b3672927 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1094.589127] env[61728]: INFO nova.compute.manager [req-c3c814e9-7de1-48e4-8eda-25b4997a1075 req-a0b4fffd-f2cb-49c0-ab00-1dfeb2a88443 service nova] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Neutron deleted interface c4f8bc0c-ff22-428b-a962-d2d8b3672927; detaching it from the instance and deleting it from the info cache [ 1094.590294] env[61728]: DEBUG nova.network.neutron [req-c3c814e9-7de1-48e4-8eda-25b4997a1075 req-a0b4fffd-f2cb-49c0-ab00-1dfeb2a88443 service nova] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.591821] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.532s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.592840] env[61728]: DEBUG nova.objects.instance [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Lazy-loading 'resources' on Instance uuid 5eb3c3fa-d612-43d8-875c-2ee595e29081 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.593870] env[61728]: DEBUG oslo_vmware.rw_handles [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lease: (returnval){ [ 1094.593870] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52590486-6f02-4ee4-a0a0-5f16ade007a2" [ 1094.593870] env[61728]: _type = "HttpNfcLease" [ 1094.593870] env[61728]: } obtained for exporting VM: (result){ [ 1094.593870] env[61728]: value = "vm-122216" [ 1094.593870] env[61728]: _type = "VirtualMachine" [ 1094.593870] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1094.594132] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the lease: (returnval){ [ 1094.594132] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52590486-6f02-4ee4-a0a0-5f16ade007a2" [ 1094.594132] env[61728]: _type = "HttpNfcLease" [ 1094.594132] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1094.603021] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1094.603021] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52590486-6f02-4ee4-a0a0-5f16ade007a2" [ 1094.603021] env[61728]: _type = "HttpNfcLease" [ 1094.603021] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1094.899024] env[61728]: INFO nova.compute.manager [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Took 18.85 seconds to build instance. [ 1094.940361] env[61728]: DEBUG nova.virt.hardware [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1094.940518] env[61728]: DEBUG nova.virt.hardware [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1094.940636] env[61728]: DEBUG nova.virt.hardware [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1094.940909] env[61728]: DEBUG nova.virt.hardware [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1094.941120] env[61728]: DEBUG nova.virt.hardware [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1094.941286] env[61728]: DEBUG nova.virt.hardware [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1094.941507] env[61728]: DEBUG nova.virt.hardware [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1094.941674] env[61728]: DEBUG nova.virt.hardware [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1094.941886] env[61728]: DEBUG nova.virt.hardware [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1094.942090] env[61728]: DEBUG nova.virt.hardware [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1094.942283] env[61728]: DEBUG nova.virt.hardware [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1094.945564] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a82e1c37-95a7-423a-a3e4-466358ce4ea1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.950771] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bdbb147a-5737-413d-81de-0827f217e7fd tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.950771] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bdbb147a-5737-413d-81de-0827f217e7fd tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquired lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.950771] env[61728]: DEBUG nova.network.neutron [None req-bdbb147a-5737-413d-81de-0827f217e7fd tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1094.950771] env[61728]: DEBUG nova.objects.instance [None req-bdbb147a-5737-413d-81de-0827f217e7fd tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lazy-loading 'info_cache' on Instance uuid d55beb02-00fc-4df6-a239-9e5d776a7c18 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.955633] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f22826c2-fa8f-441d-87b9-6750088087b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.970329] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:4c:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9732690c-bdcf-4e6f-9a32-42c196333eb8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '63b2bf95-c628-4c62-bdbd-051192e13a2f', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1094.978020] env[61728]: DEBUG oslo.service.loopingcall [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1094.979010] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1094.979256] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a94fa4d4-3a0e-40b1-8c55-1571754b8b11 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.999871] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1094.999871] env[61728]: value = "task-464903" [ 1094.999871] env[61728]: _type = "Task" [ 1094.999871] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.012926] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464903, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.057253] env[61728]: DEBUG nova.network.neutron [-] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.093088] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "f27f22aa-f05d-4981-a389-311731c51f93" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.093247] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "f27f22aa-f05d-4981-a389-311731c51f93" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.093435] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "f27f22aa-f05d-4981-a389-311731c51f93-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.093808] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "f27f22aa-f05d-4981-a389-311731c51f93-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.093927] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "f27f22aa-f05d-4981-a389-311731c51f93-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.100141] env[61728]: INFO nova.compute.manager [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Terminating instance [ 1095.101644] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c584cde2-b391-4b12-afeb-1ba429ca0b58 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.107480] env[61728]: DEBUG nova.compute.manager [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1095.107580] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1095.108533] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e012c7-2a01-422f-8926-e80c1fd22064 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.115456] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1095.115456] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52590486-6f02-4ee4-a0a0-5f16ade007a2" [ 1095.115456] env[61728]: _type = "HttpNfcLease" [ 1095.115456] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1095.118584] env[61728]: DEBUG oslo_vmware.rw_handles [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1095.118584] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52590486-6f02-4ee4-a0a0-5f16ade007a2" [ 1095.118584] env[61728]: _type = "HttpNfcLease" [ 1095.118584] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1095.119956] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1095.122165] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9caa76e1-7a96-4a28-8870-156cd99c66f7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.129166] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1520a8-287b-4f7f-9e33-4ec7912081ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.139669] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-229a46c7-57e9-4b05-902e-9a109feaf907 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.147474] env[61728]: DEBUG oslo_vmware.rw_handles [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52230a00-03e6-34fa-893e-6a300085b650/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1095.147685] env[61728]: DEBUG oslo_vmware.rw_handles [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52230a00-03e6-34fa-893e-6a300085b650/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1095.150056] env[61728]: DEBUG oslo_vmware.api [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1095.150056] env[61728]: value = "task-464904" [ 1095.150056] env[61728]: _type = "Task" [ 1095.150056] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.221712] env[61728]: DEBUG nova.compute.manager [req-c3c814e9-7de1-48e4-8eda-25b4997a1075 req-a0b4fffd-f2cb-49c0-ab00-1dfeb2a88443 service nova] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Detach interface failed, port_id=c4f8bc0c-ff22-428b-a962-d2d8b3672927, reason: Instance 32a781b0-2db4-4883-ae96-d94c8181902b could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1095.227120] env[61728]: DEBUG oslo_vmware.api [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464904, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.258566] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-134a4008-eafe-4ed9-9057-0bda81bdef0c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.371895] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7721dcac-390c-4b3d-918d-5e6b1c99ae70 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.380419] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4591984b-1a91-4a23-b7a8-938361d92bc4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.409900] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2897825c-7fb6-4f22-b385-1ecff607335e tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "8ea37e93-16fc-4c60-9949-17656218d46a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.374s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.410949] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "8ea37e93-16fc-4c60-9949-17656218d46a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.221s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.410949] env[61728]: INFO nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] During sync_power_state the instance has a pending task (spawning). Skip. [ 1095.410949] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "8ea37e93-16fc-4c60-9949-17656218d46a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.411591] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27485220-3a72-47e4-844f-5b6e1b256bce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.420431] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b2d2fb-21e9-4288-ad07-bdd452186653 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.436704] env[61728]: DEBUG nova.compute.provider_tree [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.452154] env[61728]: DEBUG nova.objects.base [None req-bdbb147a-5737-413d-81de-0827f217e7fd tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1096.263660] env[61728]: INFO nova.compute.manager [-] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Took 1.97 seconds to deallocate network for instance. [ 1096.264562] env[61728]: DEBUG nova.scheduler.client.report [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1096.290806] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464903, 'name': CreateVM_Task, 'duration_secs': 0.441565} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.293802] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1096.294342] env[61728]: DEBUG oslo_vmware.api [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464904, 'name': PowerOffVM_Task, 'duration_secs': 0.268803} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.295596] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.295596] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.296112] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1096.296475] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1096.296726] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1096.297389] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb72f22d-415b-451f-ba89-03a779ab3d7f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.300495] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c03622c8-4ab3-4e24-af29-91203608b128 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.306264] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1096.306264] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524871ad-4364-587b-5da4-9bcfc09ddc8e" [ 1096.306264] env[61728]: _type = "Task" [ 1096.306264] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.315113] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524871ad-4364-587b-5da4-9bcfc09ddc8e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.366557] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1096.366885] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1096.367277] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Deleting the datastore file [datastore1] f27f22aa-f05d-4981-a389-311731c51f93 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1096.367625] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f11ecfd-1ffa-4d89-a65d-f995f950e482 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.377147] env[61728]: DEBUG oslo_vmware.api [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for the task: (returnval){ [ 1096.377147] env[61728]: value = "task-464906" [ 1096.377147] env[61728]: _type = "Task" [ 1096.377147] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.386903] env[61728]: DEBUG oslo_vmware.api [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464906, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.529647] env[61728]: DEBUG nova.network.neutron [None req-bdbb147a-5737-413d-81de-0827f217e7fd tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Updating instance_info_cache with network_info: [{"id": "4155b1a7-975f-404c-8374-47665c8046ca", "address": "fa:16:3e:e4:b3:9b", "network": {"id": "72d11e2a-907e-419e-b8a4-53aa39f2566a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1755247352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833a646ea21a48c4951ed8cedebaa98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3836fb52-19c6-4e10-a0ca-f0bca73dc887", "external-id": "nsx-vlan-transportzone-964", "segmentation_id": 964, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4155b1a7-97", "ovs_interfaceid": "4155b1a7-975f-404c-8374-47665c8046ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.776636] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.184s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.783711] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b1fd2b96-6392-44b0-b023-a9004cc0f285 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "8ea37e93-16fc-4c60-9949-17656218d46a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.783882] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b1fd2b96-6392-44b0-b023-a9004cc0f285 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "8ea37e93-16fc-4c60-9949-17656218d46a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.784101] env[61728]: DEBUG nova.compute.manager [None req-b1fd2b96-6392-44b0-b023-a9004cc0f285 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1096.785714] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.785972] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.786174] env[61728]: DEBUG nova.objects.instance [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lazy-loading 'resources' on Instance uuid 32a781b0-2db4-4883-ae96-d94c8181902b {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.787943] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c610e0-e5d3-44b9-bb1f-6a90ee3dc5fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.796416] env[61728]: DEBUG nova.compute.manager [None req-b1fd2b96-6392-44b0-b023-a9004cc0f285 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61728) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1096.797440] env[61728]: DEBUG nova.objects.instance [None req-b1fd2b96-6392-44b0-b023-a9004cc0f285 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lazy-loading 'flavor' on Instance uuid 8ea37e93-16fc-4c60-9949-17656218d46a {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.813843] env[61728]: INFO nova.scheduler.client.report [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Deleted allocations for instance 5eb3c3fa-d612-43d8-875c-2ee595e29081 [ 1096.825677] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524871ad-4364-587b-5da4-9bcfc09ddc8e, 'name': SearchDatastore_Task, 'duration_secs': 0.012347} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.825795] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.826023] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1096.826769] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.826769] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.826769] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1096.826958] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e89a3620-08a0-4aa5-b107-879f938f372d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.837102] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1096.837102] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1096.837224] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9449618-1699-470a-b036-bead9166e299 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.844212] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1096.844212] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d65336-3eb9-6efd-bd92-86bb9a894443" [ 1096.844212] env[61728]: _type = "Task" [ 1096.844212] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.852666] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d65336-3eb9-6efd-bd92-86bb9a894443, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.889108] env[61728]: DEBUG oslo_vmware.api [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Task: {'id': task-464906, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.227811} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.889485] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1096.889745] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1096.890446] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1096.890446] env[61728]: INFO nova.compute.manager [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Took 1.78 seconds to destroy the instance on the hypervisor. [ 1096.890595] env[61728]: DEBUG oslo.service.loopingcall [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1096.890889] env[61728]: DEBUG nova.compute.manager [-] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1096.890998] env[61728]: DEBUG nova.network.neutron [-] [instance: f27f22aa-f05d-4981-a389-311731c51f93] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1097.033111] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bdbb147a-5737-413d-81de-0827f217e7fd tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Releasing lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.305632] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1fd2b96-6392-44b0-b023-a9004cc0f285 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1097.305632] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35e70a00-b785-4d5b-97da-709c431719be {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.313453] env[61728]: DEBUG oslo_vmware.api [None req-b1fd2b96-6392-44b0-b023-a9004cc0f285 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1097.313453] env[61728]: value = "task-464907" [ 1097.313453] env[61728]: _type = "Task" [ 1097.313453] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.325723] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f2ba09be-a8cc-4ff6-b6f6-a056aa5902a5 tempest-InstanceActionsV221TestJSON-876924873 tempest-InstanceActionsV221TestJSON-876924873-project-member] Lock "5eb3c3fa-d612-43d8-875c-2ee595e29081" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.396s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.326928] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "5eb3c3fa-d612-43d8-875c-2ee595e29081" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 6.138s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.327182] env[61728]: INFO nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] During sync_power_state the instance has a pending task (deleting). Skip. [ 1097.327441] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "5eb3c3fa-d612-43d8-875c-2ee595e29081" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.336023] env[61728]: DEBUG oslo_vmware.api [None req-b1fd2b96-6392-44b0-b023-a9004cc0f285 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464907, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.362262] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d65336-3eb9-6efd-bd92-86bb9a894443, 'name': SearchDatastore_Task, 'duration_secs': 0.009909} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.363197] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbbb8c2e-bf3d-43a7-9020-b8ea0aa1d2c9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.377791] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1097.377791] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f0e177-fb9d-00a2-e1a9-30643bf9e8fd" [ 1097.377791] env[61728]: _type = "Task" [ 1097.377791] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.395252] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f0e177-fb9d-00a2-e1a9-30643bf9e8fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.395252] env[61728]: DEBUG nova.compute.manager [req-d1710053-a511-4299-af2f-c94e89b06618 req-e7e7e766-f4aa-40e6-a084-4811877c18f4 service nova] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Received event network-vif-deleted-4a6313df-8741-4533-ae0e-1f469193c389 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1097.395252] env[61728]: INFO nova.compute.manager [req-d1710053-a511-4299-af2f-c94e89b06618 req-e7e7e766-f4aa-40e6-a084-4811877c18f4 service nova] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Neutron deleted interface 4a6313df-8741-4533-ae0e-1f469193c389; detaching it from the instance and deleting it from the info cache [ 1097.395252] env[61728]: DEBUG nova.network.neutron [req-d1710053-a511-4299-af2f-c94e89b06618 req-e7e7e766-f4aa-40e6-a084-4811877c18f4 service nova] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.527073] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7810e1-f573-42f1-9a05-ec11246b2935 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.535877] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2315ad94-6859-4265-8153-c7203d812ffc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.540679] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdbb147a-5737-413d-81de-0827f217e7fd tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1097.541095] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90b79d2b-1802-4c91-a4e2-04d5d9f4474f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.577792] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d4c94a-199d-4480-974d-1011a9e3e5a1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.580912] env[61728]: DEBUG oslo_vmware.api [None req-bdbb147a-5737-413d-81de-0827f217e7fd tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1097.580912] env[61728]: value = "task-464908" [ 1097.580912] env[61728]: _type = "Task" [ 1097.580912] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.590776] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53ce9bb-69eb-42fa-a270-efc0e21c3060 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.600629] env[61728]: DEBUG oslo_vmware.api [None req-bdbb147a-5737-413d-81de-0827f217e7fd tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464908, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.612147] env[61728]: DEBUG nova.compute.provider_tree [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1097.779093] env[61728]: DEBUG nova.network.neutron [-] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.824950] env[61728]: DEBUG oslo_vmware.api [None req-b1fd2b96-6392-44b0-b023-a9004cc0f285 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464907, 'name': PowerOffVM_Task, 'duration_secs': 0.228935} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.825325] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1fd2b96-6392-44b0-b023-a9004cc0f285 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1097.825473] env[61728]: DEBUG nova.compute.manager [None req-b1fd2b96-6392-44b0-b023-a9004cc0f285 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1097.826916] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8493e3-8dc7-40ef-bcc0-e96439139778 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.890147] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f0e177-fb9d-00a2-e1a9-30643bf9e8fd, 'name': SearchDatastore_Task, 'duration_secs': 0.012159} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.890470] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.890783] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 41329d0d-3838-4d1b-a4e6-8df4508eea76/41329d0d-3838-4d1b-a4e6-8df4508eea76.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1097.891082] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-000196b6-e4b0-4786-905b-293a8a6ef665 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.899173] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-59a41724-faac-42ac-bc61-5e984a75b767 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.901452] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1097.901452] env[61728]: value = "task-464909" [ 1097.901452] env[61728]: _type = "Task" [ 1097.901452] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.909951] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6df96c6f-a408-4e31-93b4-8928b4a2484b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.925768] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464909, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.947981] env[61728]: DEBUG nova.compute.manager [req-d1710053-a511-4299-af2f-c94e89b06618 req-e7e7e766-f4aa-40e6-a084-4811877c18f4 service nova] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Detach interface failed, port_id=4a6313df-8741-4533-ae0e-1f469193c389, reason: Instance f27f22aa-f05d-4981-a389-311731c51f93 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1098.093190] env[61728]: DEBUG oslo_vmware.api [None req-bdbb147a-5737-413d-81de-0827f217e7fd tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464908, 'name': PowerOnVM_Task, 'duration_secs': 0.433172} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.093517] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdbb147a-5737-413d-81de-0827f217e7fd tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1098.093728] env[61728]: DEBUG nova.compute.manager [None req-bdbb147a-5737-413d-81de-0827f217e7fd tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1098.094558] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a9d9303-4bb1-4d40-bf20-28406c4a698f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.145565] env[61728]: ERROR nova.scheduler.client.report [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [req-5be94e7d-1a7b-4ef1-8077-74608f2528a4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5be94e7d-1a7b-4ef1-8077-74608f2528a4"}]} [ 1098.163935] env[61728]: DEBUG nova.scheduler.client.report [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1098.193188] env[61728]: DEBUG nova.scheduler.client.report [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1098.193434] env[61728]: DEBUG nova.compute.provider_tree [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1098.212580] env[61728]: DEBUG nova.scheduler.client.report [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1098.235029] env[61728]: DEBUG nova.scheduler.client.report [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1098.281585] env[61728]: INFO nova.compute.manager [-] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Took 1.39 seconds to deallocate network for instance. [ 1098.344765] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b1fd2b96-6392-44b0-b023-a9004cc0f285 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "8ea37e93-16fc-4c60-9949-17656218d46a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.561s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.412419] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464909, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.455778] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd8e14b-8b54-4dad-aa11-84075103ea25 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.466660] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba8473b-3f56-4910-9c83-d0ed5daebd9e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.500506] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb151f3d-0637-479b-ae5a-07b444f6d368 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.508460] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c301581-912f-4183-8fbe-611e2d07bcaf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.523649] env[61728]: DEBUG nova.compute.provider_tree [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1098.792179] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.914388] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464909, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.61307} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.914686] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 41329d0d-3838-4d1b-a4e6-8df4508eea76/41329d0d-3838-4d1b-a4e6-8df4508eea76.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1098.914912] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1098.915200] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4f239668-b5e8-4b2e-b4b4-006680d4a7ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.924656] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1098.924656] env[61728]: value = "task-464910" [ 1098.924656] env[61728]: _type = "Task" [ 1098.924656] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.939683] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464910, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.064301] env[61728]: DEBUG nova.scheduler.client.report [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 173 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1099.064500] env[61728]: DEBUG nova.compute.provider_tree [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 173 to 174 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1099.064688] env[61728]: DEBUG nova.compute.provider_tree [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1099.434821] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464910, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079533} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.435129] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1099.435934] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b07db9-c944-4874-91ac-c0b537bba3f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.457833] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 41329d0d-3838-4d1b-a4e6-8df4508eea76/41329d0d-3838-4d1b-a4e6-8df4508eea76.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1099.458343] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "8ea37e93-16fc-4c60-9949-17656218d46a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.458596] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "8ea37e93-16fc-4c60-9949-17656218d46a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.459083] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "8ea37e93-16fc-4c60-9949-17656218d46a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.459140] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "8ea37e93-16fc-4c60-9949-17656218d46a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.460057] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "8ea37e93-16fc-4c60-9949-17656218d46a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.461772] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9b0521b-b1a2-4eb0-bb0b-60e588a447e6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.478505] env[61728]: INFO nova.compute.manager [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Terminating instance [ 1099.481291] env[61728]: DEBUG nova.compute.manager [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1099.481855] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1099.482807] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371920a9-d8f5-461c-a81e-8b27a91c43d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.487624] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1099.487624] env[61728]: value = "task-464911" [ 1099.487624] env[61728]: _type = "Task" [ 1099.487624] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.489750] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1099.492945] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2bffa751-f1bf-4c13-abf3-bb9062441a3b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.500379] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464911, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.575602] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.788s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.576717] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.785s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.577198] env[61728]: DEBUG nova.objects.instance [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lazy-loading 'resources' on Instance uuid f27f22aa-f05d-4981-a389-311731c51f93 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.583181] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1099.583181] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1099.583181] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleting the datastore file [datastore1] 8ea37e93-16fc-4c60-9949-17656218d46a {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1099.583181] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a51e830a-b0fa-47c6-8498-f07d9e51b284 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.589687] env[61728]: DEBUG oslo_vmware.api [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1099.589687] env[61728]: value = "task-464913" [ 1099.589687] env[61728]: _type = "Task" [ 1099.589687] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.599392] env[61728]: DEBUG oslo_vmware.api [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464913, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.610084] env[61728]: INFO nova.scheduler.client.report [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleted allocations for instance 32a781b0-2db4-4883-ae96-d94c8181902b [ 1100.003052] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464911, 'name': ReconfigVM_Task, 'duration_secs': 0.300029} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.003494] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 41329d0d-3838-4d1b-a4e6-8df4508eea76/41329d0d-3838-4d1b-a4e6-8df4508eea76.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1100.005012] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_options': None, 'guest_format': None, 'encryption_secret_uuid': None, 'disk_bus': None, 'device_name': '/dev/sda', 'encryption_format': None, 'boot_index': 0, 'encrypted': False, 'size': 0, 'device_type': 'disk', 'image_id': '8b767102-1435-4827-a43b-8e2e25ec780b'}], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'f3a466d5-7e2e-4552-8d87-620dad31cf0e', 'delete_on_termination': False, 'guest_format': None, 'disk_bus': None, 'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122206', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'name': 'volume-76406f6c-49da-4712-8faa-8af3a2f760a1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41329d0d-3838-4d1b-a4e6-8df4508eea76', 'attached_at': '', 'detached_at': '', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'serial': '76406f6c-49da-4712-8faa-8af3a2f760a1'}, 'mount_device': '/dev/sdb', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=61728) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1100.005270] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1100.005494] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122206', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'name': 'volume-76406f6c-49da-4712-8faa-8af3a2f760a1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41329d0d-3838-4d1b-a4e6-8df4508eea76', 'attached_at': '', 'detached_at': '', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'serial': '76406f6c-49da-4712-8faa-8af3a2f760a1'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1100.006450] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec61c207-5095-40e9-96f9-1782041b78a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.030956] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a95271-16d3-46ed-ae28-f68516443f99 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.057874] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] volume-76406f6c-49da-4712-8faa-8af3a2f760a1/volume-76406f6c-49da-4712-8faa-8af3a2f760a1.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1100.058292] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f10f5e9e-a229-4b4f-b1f8-339cca7a828f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.076659] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1100.076659] env[61728]: value = "task-464914" [ 1100.076659] env[61728]: _type = "Task" [ 1100.076659] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.089494] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464914, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.100903] env[61728]: DEBUG oslo_vmware.api [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464913, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198106} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.101172] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1100.101366] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1100.101621] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1100.101763] env[61728]: INFO nova.compute.manager [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1100.102126] env[61728]: DEBUG oslo.service.loopingcall [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1100.102246] env[61728]: DEBUG nova.compute.manager [-] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1100.102346] env[61728]: DEBUG nova.network.neutron [-] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1100.120396] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cfa9e67f-1f0c-49f7-b7e2-feccf23d0eac tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "32a781b0-2db4-4883-ae96-d94c8181902b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.957s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.248824] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3001418-ed57-402f-908b-9804e683b535 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.259011] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deb0e08a-812e-4033-bb95-fdcea55f8c0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.293693] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c750b67-ba8b-416d-8103-6b11943c97e5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.303291] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68eb87f5-acb1-4e13-bad7-e83050cea0ef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.317265] env[61728]: DEBUG nova.compute.provider_tree [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1100.461201] env[61728]: DEBUG nova.compute.manager [req-f0a92544-d404-4338-98c5-88aa97242bc2 req-6fd1cac5-bf33-421e-ac1e-5202cadd524b service nova] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Received event network-vif-deleted-ce80d456-9e6c-4d02-a7c6-99593a0761c6 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1100.461201] env[61728]: INFO nova.compute.manager [req-f0a92544-d404-4338-98c5-88aa97242bc2 req-6fd1cac5-bf33-421e-ac1e-5202cadd524b service nova] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Neutron deleted interface ce80d456-9e6c-4d02-a7c6-99593a0761c6; detaching it from the instance and deleting it from the info cache [ 1100.461201] env[61728]: DEBUG nova.network.neutron [req-f0a92544-d404-4338-98c5-88aa97242bc2 req-6fd1cac5-bf33-421e-ac1e-5202cadd524b service nova] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.588474] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464914, 'name': ReconfigVM_Task, 'duration_secs': 0.30012} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.588474] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Reconfigured VM instance instance-0000006b to attach disk [datastore1] volume-76406f6c-49da-4712-8faa-8af3a2f760a1/volume-76406f6c-49da-4712-8faa-8af3a2f760a1.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1100.593687] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24d5a120-ec3e-4d09-905c-129d827e31f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.609727] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1100.609727] env[61728]: value = "task-464915" [ 1100.609727] env[61728]: _type = "Task" [ 1100.609727] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.619428] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464915, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.826055] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "79782d2b-1b90-4f3f-a1ce-394e88a00546" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.826055] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "79782d2b-1b90-4f3f-a1ce-394e88a00546" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.845728] env[61728]: ERROR nova.scheduler.client.report [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [req-ee30e2e5-ab6a-49d6-b1e6-803cd7e65c57] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ee30e2e5-ab6a-49d6-b1e6-803cd7e65c57"}]} [ 1100.864438] env[61728]: DEBUG nova.scheduler.client.report [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1100.879765] env[61728]: DEBUG nova.scheduler.client.report [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1100.880072] env[61728]: DEBUG nova.compute.provider_tree [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1100.894423] env[61728]: DEBUG nova.scheduler.client.report [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1100.920873] env[61728]: DEBUG nova.scheduler.client.report [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1100.936771] env[61728]: DEBUG nova.network.neutron [-] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.964617] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3e2f0a94-7ddf-44f6-a698-c744bfd1f24a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.976921] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d8c6274-4cc5-4b17-9c43-0aecc51ba9be {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.014334] env[61728]: DEBUG nova.compute.manager [req-f0a92544-d404-4338-98c5-88aa97242bc2 req-6fd1cac5-bf33-421e-ac1e-5202cadd524b service nova] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Detach interface failed, port_id=ce80d456-9e6c-4d02-a7c6-99593a0761c6, reason: Instance 8ea37e93-16fc-4c60-9949-17656218d46a could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1101.115838] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f713dc-2920-44d7-9f6b-056bae33e7c9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.121157] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464915, 'name': ReconfigVM_Task, 'duration_secs': 0.258578} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.121836] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122206', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'name': 'volume-76406f6c-49da-4712-8faa-8af3a2f760a1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41329d0d-3838-4d1b-a4e6-8df4508eea76', 'attached_at': '', 'detached_at': '', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'serial': '76406f6c-49da-4712-8faa-8af3a2f760a1'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1101.122424] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ed65b866-2d5a-47ad-95ac-af0c7df2d23b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.126458] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58899746-e44f-49f2-b455-a41ce65df4c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.130734] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1101.130734] env[61728]: value = "task-464916" [ 1101.130734] env[61728]: _type = "Task" [ 1101.130734] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.162173] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70226209-135d-45ea-92b6-a16a14dc4fa2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.168691] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464916, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.174611] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dcd9004-ab05-4bd8-88c7-0e0d25ccf67c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.187765] env[61728]: DEBUG nova.compute.provider_tree [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1101.331037] env[61728]: DEBUG nova.compute.manager [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1101.439435] env[61728]: INFO nova.compute.manager [-] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Took 1.34 seconds to deallocate network for instance. [ 1101.643065] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464916, 'name': Rename_Task, 'duration_secs': 0.235501} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.643467] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1101.643467] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f0d2801-dfe3-4e42-b8c5-f0001963b2c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.649816] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1101.649816] env[61728]: value = "task-464917" [ 1101.649816] env[61728]: _type = "Task" [ 1101.649816] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.657193] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464917, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.710402] env[61728]: ERROR nova.scheduler.client.report [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] [req-f8f5222b-c17a-4efd-8224-3d695c9f5f2b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f8f5222b-c17a-4efd-8224-3d695c9f5f2b"}]} [ 1101.726316] env[61728]: DEBUG nova.scheduler.client.report [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1101.739014] env[61728]: DEBUG nova.scheduler.client.report [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1101.739263] env[61728]: DEBUG nova.compute.provider_tree [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1101.749891] env[61728]: DEBUG nova.scheduler.client.report [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1101.765865] env[61728]: DEBUG nova.scheduler.client.report [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1101.850591] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.928313] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d8bace-1564-442d-b426-37e9bf1ff74d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.936823] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14eff0be-3f4e-4bd4-a85b-7bac2bbc8552 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.977364] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.977364] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d695d21-01f6-403c-ae92-0635215807e8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.985529] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cdbcd94-0aa5-4229-9623-a40a2ae04293 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.006161] env[61728]: DEBUG nova.compute.provider_tree [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1102.161755] env[61728]: DEBUG oslo_vmware.api [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464917, 'name': PowerOnVM_Task, 'duration_secs': 0.467586} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.162149] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1102.162386] env[61728]: DEBUG nova.compute.manager [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1102.163154] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d852914c-6fec-44ce-835a-8416c0ad8df3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.382184] env[61728]: DEBUG nova.compute.manager [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Stashing vm_state: active {{(pid=61728) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1102.555675] env[61728]: DEBUG nova.scheduler.client.report [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 177 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1102.556014] env[61728]: DEBUG nova.compute.provider_tree [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 177 to 178 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1102.556277] env[61728]: DEBUG nova.compute.provider_tree [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1102.680118] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.904290] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.063060] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.485s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.065035] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.214s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.066105] env[61728]: INFO nova.compute.claims [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1103.080027] env[61728]: INFO nova.scheduler.client.report [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Deleted allocations for instance f27f22aa-f05d-4981-a389-311731c51f93 [ 1103.587446] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e85b536c-98e2-4b6c-823b-bda3066cca74 tempest-ServerRescueNegativeTestJSON-748358644 tempest-ServerRescueNegativeTestJSON-748358644-project-member] Lock "f27f22aa-f05d-4981-a389-311731c51f93" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.494s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.145674] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.145953] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.234664] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f30b9c8-a763-4999-aa96-23da25b98a79 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.242943] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb9d595-5190-41e7-a633-ccb73e8a5f37 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.274645] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51cf881-6cd2-4182-9475-39569b3711b7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.282491] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7548356-fab2-49af-8bd5-883a114b1ddb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.296410] env[61728]: DEBUG nova.compute.provider_tree [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.576394] env[61728]: DEBUG oslo_vmware.rw_handles [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52230a00-03e6-34fa-893e-6a300085b650/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1104.577355] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ebb81b8-8391-45d7-80ae-9c5f73674133 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.583933] env[61728]: DEBUG oslo_vmware.rw_handles [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52230a00-03e6-34fa-893e-6a300085b650/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1104.584135] env[61728]: ERROR oslo_vmware.rw_handles [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52230a00-03e6-34fa-893e-6a300085b650/disk-0.vmdk due to incomplete transfer. [ 1104.584369] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-21cc3bce-5e2b-486c-b780-82fa74e5334a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.591423] env[61728]: DEBUG oslo_vmware.rw_handles [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52230a00-03e6-34fa-893e-6a300085b650/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1104.591623] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Uploaded image d9914484-11b1-4819-982a-67a108bd3ed1 to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1104.594180] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1104.594426] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f41dbe94-4c37-4911-b300-8b433a516401 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.600698] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1104.600698] env[61728]: value = "task-464918" [ 1104.600698] env[61728]: _type = "Task" [ 1104.600698] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.607920] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464918, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.650032] env[61728]: INFO nova.compute.manager [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Detaching volume 76406f6c-49da-4712-8faa-8af3a2f760a1 [ 1104.693264] env[61728]: INFO nova.virt.block_device [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Attempting to driver detach volume 76406f6c-49da-4712-8faa-8af3a2f760a1 from mountpoint /dev/sdb [ 1104.693702] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1104.693702] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122206', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'name': 'volume-76406f6c-49da-4712-8faa-8af3a2f760a1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41329d0d-3838-4d1b-a4e6-8df4508eea76', 'attached_at': '', 'detached_at': '', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'serial': '76406f6c-49da-4712-8faa-8af3a2f760a1'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1104.694669] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d008b040-98bd-4e11-8d16-a115f64f47ef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.725800] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb23390-a106-4950-9a3e-25394d3ac32a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.733459] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499121c0-a89d-4c47-935f-c8e7af275d16 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.755087] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ff85c2-71ee-4a6e-a037-a6fb779f40a3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.771951] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] The volume has not been displaced from its original location: [datastore1] volume-76406f6c-49da-4712-8faa-8af3a2f760a1/volume-76406f6c-49da-4712-8faa-8af3a2f760a1.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1104.778223] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1104.778602] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c4ea10a-11e5-4719-a74d-9fa8546d5d90 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.796909] env[61728]: DEBUG oslo_vmware.api [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1104.796909] env[61728]: value = "task-464919" [ 1104.796909] env[61728]: _type = "Task" [ 1104.796909] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.803790] env[61728]: DEBUG nova.scheduler.client.report [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1104.813679] env[61728]: DEBUG oslo_vmware.api [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464919, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.110286] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464918, 'name': Destroy_Task, 'duration_secs': 0.334877} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.110600] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Destroyed the VM [ 1105.110846] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1105.111122] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4ca64ebb-88c9-4151-94bd-bf8fc17d6aa6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.117412] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1105.117412] env[61728]: value = "task-464920" [ 1105.117412] env[61728]: _type = "Task" [ 1105.117412] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.126496] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464920, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.307506] env[61728]: DEBUG oslo_vmware.api [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464919, 'name': ReconfigVM_Task, 'duration_secs': 0.45554} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.307861] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1105.313129] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.249s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.313680] env[61728]: DEBUG nova.compute.manager [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1105.316336] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73a40835-5c03-4054-bd18-6ab07547efcf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.327018] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.354s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.327338] env[61728]: DEBUG nova.objects.instance [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lazy-loading 'resources' on Instance uuid 8ea37e93-16fc-4c60-9949-17656218d46a {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.333458] env[61728]: DEBUG oslo_vmware.api [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1105.333458] env[61728]: value = "task-464921" [ 1105.333458] env[61728]: _type = "Task" [ 1105.333458] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.343024] env[61728]: DEBUG oslo_vmware.api [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464921, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.627924] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464920, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.828573] env[61728]: DEBUG nova.compute.utils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1105.830018] env[61728]: DEBUG nova.compute.manager [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1105.830207] env[61728]: DEBUG nova.network.neutron [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1105.844408] env[61728]: DEBUG oslo_vmware.api [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464921, 'name': ReconfigVM_Task, 'duration_secs': 0.154683} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.844735] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122206', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'name': 'volume-76406f6c-49da-4712-8faa-8af3a2f760a1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41329d0d-3838-4d1b-a4e6-8df4508eea76', 'attached_at': '', 'detached_at': '', 'volume_id': '76406f6c-49da-4712-8faa-8af3a2f760a1', 'serial': '76406f6c-49da-4712-8faa-8af3a2f760a1'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1105.883198] env[61728]: DEBUG nova.policy [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '39df6cef3a6f430d92e15d68076e759f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '909c86730f554a04ab05dcad3a53e9dd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1106.005550] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee26924-bd4d-47db-a3e7-06b0a49e89ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.014550] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-188e4986-fe2c-489f-8f91-dd634126e432 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.050845] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8332e355-2bb4-4104-9d65-72a7e304ae09 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.060614] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3907c85-dfd2-4314-8884-4fe6909a77a0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.074843] env[61728]: DEBUG nova.compute.provider_tree [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.128540] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464920, 'name': RemoveSnapshot_Task, 'duration_secs': 0.771995} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.128540] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1106.128540] env[61728]: DEBUG nova.compute.manager [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1106.129023] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0b7b22-cb96-4d08-95f2-f257c603dbcb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.168501] env[61728]: DEBUG nova.network.neutron [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Successfully created port: 28c67e6e-1201-4072-8c7d-331a803bada0 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1106.338019] env[61728]: DEBUG nova.compute.manager [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1106.399113] env[61728]: DEBUG nova.objects.instance [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lazy-loading 'flavor' on Instance uuid 41329d0d-3838-4d1b-a4e6-8df4508eea76 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1106.412934] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.578316] env[61728]: DEBUG nova.scheduler.client.report [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1106.642344] env[61728]: INFO nova.compute.manager [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Shelve offloading [ 1106.644375] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1106.644644] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5152e3cb-e949-4a22-a9dc-9846a60112bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.652936] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1106.652936] env[61728]: value = "task-464922" [ 1106.652936] env[61728]: _type = "Task" [ 1106.652936] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.661938] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1106.662175] env[61728]: DEBUG nova.compute.manager [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1106.662925] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-837a328a-8f6b-4721-a775-3c52225f6165 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.668589] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.668734] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.669051] env[61728]: DEBUG nova.network.neutron [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1106.727322] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquiring lock "dbe877e6-1754-4cf7-b454-abd88bfa713a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.727604] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Lock "dbe877e6-1754-4cf7-b454-abd88bfa713a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.854980] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.083525] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.756s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.086027] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 4.406s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.086167] env[61728]: DEBUG nova.objects.instance [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61728) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1107.104267] env[61728]: INFO nova.scheduler.client.report [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleted allocations for instance 8ea37e93-16fc-4c60-9949-17656218d46a [ 1107.230026] env[61728]: DEBUG nova.compute.manager [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1107.344625] env[61728]: DEBUG nova.compute.manager [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1107.370049] env[61728]: DEBUG nova.virt.hardware [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1107.370328] env[61728]: DEBUG nova.virt.hardware [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1107.370494] env[61728]: DEBUG nova.virt.hardware [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1107.370684] env[61728]: DEBUG nova.virt.hardware [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1107.370838] env[61728]: DEBUG nova.virt.hardware [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1107.370996] env[61728]: DEBUG nova.virt.hardware [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1107.371237] env[61728]: DEBUG nova.virt.hardware [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1107.371402] env[61728]: DEBUG nova.virt.hardware [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1107.371575] env[61728]: DEBUG nova.virt.hardware [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1107.371779] env[61728]: DEBUG nova.virt.hardware [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1107.371966] env[61728]: DEBUG nova.virt.hardware [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1107.372877] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bdb713e-4f68-418c-91f7-e82a0c7d5c5e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.380564] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f60167b-9318-46fa-b09d-6d528f292556 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.394188] env[61728]: DEBUG nova.network.neutron [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Updating instance_info_cache with network_info: [{"id": "ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb", "address": "fa:16:3e:36:bc:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea6ae87d-d3", "ovs_interfaceid": "ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.405735] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ff975aab-9ba2-42ce-8bf4-e9be2a9a3eb0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.260s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.610667] env[61728]: DEBUG oslo_concurrency.lockutils [None req-02cb2099-760d-4163-b994-d0869812d378 tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "8ea37e93-16fc-4c60-9949-17656218d46a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.152s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.629619] env[61728]: DEBUG nova.compute.manager [req-c8ad1688-44d3-466c-9ad5-60791961c675 req-fdda8b1c-cbe0-4b47-8c10-7d85b2f8155f service nova] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Received event network-vif-plugged-28c67e6e-1201-4072-8c7d-331a803bada0 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1107.629937] env[61728]: DEBUG oslo_concurrency.lockutils [req-c8ad1688-44d3-466c-9ad5-60791961c675 req-fdda8b1c-cbe0-4b47-8c10-7d85b2f8155f service nova] Acquiring lock "79782d2b-1b90-4f3f-a1ce-394e88a00546-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.630135] env[61728]: DEBUG oslo_concurrency.lockutils [req-c8ad1688-44d3-466c-9ad5-60791961c675 req-fdda8b1c-cbe0-4b47-8c10-7d85b2f8155f service nova] Lock "79782d2b-1b90-4f3f-a1ce-394e88a00546-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.630323] env[61728]: DEBUG oslo_concurrency.lockutils [req-c8ad1688-44d3-466c-9ad5-60791961c675 req-fdda8b1c-cbe0-4b47-8c10-7d85b2f8155f service nova] Lock "79782d2b-1b90-4f3f-a1ce-394e88a00546-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.630557] env[61728]: DEBUG nova.compute.manager [req-c8ad1688-44d3-466c-9ad5-60791961c675 req-fdda8b1c-cbe0-4b47-8c10-7d85b2f8155f service nova] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] No waiting events found dispatching network-vif-plugged-28c67e6e-1201-4072-8c7d-331a803bada0 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1107.630766] env[61728]: WARNING nova.compute.manager [req-c8ad1688-44d3-466c-9ad5-60791961c675 req-fdda8b1c-cbe0-4b47-8c10-7d85b2f8155f service nova] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Received unexpected event network-vif-plugged-28c67e6e-1201-4072-8c7d-331a803bada0 for instance with vm_state building and task_state spawning. [ 1107.684052] env[61728]: DEBUG nova.network.neutron [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Successfully updated port: 28c67e6e-1201-4072-8c7d-331a803bada0 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1107.748693] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.855079] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.855959] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Starting heal instance info cache {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1107.896511] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.094985] env[61728]: DEBUG oslo_concurrency.lockutils [None req-50c5bbc2-151e-45bc-ac87-09f91b283c62 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.096117] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.192s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.186833] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "refresh_cache-79782d2b-1b90-4f3f-a1ce-394e88a00546" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.187109] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquired lock "refresh_cache-79782d2b-1b90-4f3f-a1ce-394e88a00546" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.187283] env[61728]: DEBUG nova.network.neutron [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1108.431945] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1108.432865] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c814d8-1aa9-4306-844c-ba02197cface {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.440862] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1108.441336] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b245304f-5ec3-4f97-ab9c-389393985408 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.447243] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.447466] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.447675] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "41329d0d-3838-4d1b-a4e6-8df4508eea76-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.447920] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.448110] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.450163] env[61728]: INFO nova.compute.manager [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Terminating instance [ 1108.452101] env[61728]: DEBUG nova.compute.manager [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1108.452299] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1108.452993] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f795a949-46eb-4c20-89cd-2999781c2b57 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.459547] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1108.459768] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-222af7fc-86f1-4131-afe8-129a87e4e9c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.465105] env[61728]: DEBUG oslo_vmware.api [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1108.465105] env[61728]: value = "task-464924" [ 1108.465105] env[61728]: _type = "Task" [ 1108.465105] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.472387] env[61728]: DEBUG oslo_vmware.api [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464924, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.507102] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1108.507284] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1108.507477] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleting the datastore file [datastore1] 77b8f509-17e7-4c25-a43d-1655c16ce8d1 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1108.507751] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df6258c4-8f3d-4c86-830c-f449fdccef1e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.514189] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1108.514189] env[61728]: value = "task-464925" [ 1108.514189] env[61728]: _type = "Task" [ 1108.514189] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.521698] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464925, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.601509] env[61728]: INFO nova.compute.claims [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1108.717334] env[61728]: DEBUG nova.network.neutron [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1108.799288] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "40ef29df-12d6-4274-846a-4a2c32752b8b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.799704] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "40ef29df-12d6-4274-846a-4a2c32752b8b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.800065] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "40ef29df-12d6-4274-846a-4a2c32752b8b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.800390] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "40ef29df-12d6-4274-846a-4a2c32752b8b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.800685] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "40ef29df-12d6-4274-846a-4a2c32752b8b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.802965] env[61728]: INFO nova.compute.manager [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Terminating instance [ 1108.804685] env[61728]: DEBUG nova.compute.manager [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1108.804896] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1108.805716] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4baf86-064b-48cc-a335-c8b4ce76f56d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.813211] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1108.813450] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-948d03b4-7723-41b5-9f27-e6551b613774 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.819500] env[61728]: DEBUG oslo_vmware.api [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1108.819500] env[61728]: value = "task-464926" [ 1108.819500] env[61728]: _type = "Task" [ 1108.819500] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.822884] env[61728]: DEBUG nova.network.neutron [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Updating instance_info_cache with network_info: [{"id": "28c67e6e-1201-4072-8c7d-331a803bada0", "address": "fa:16:3e:32:c1:e0", "network": {"id": "4bb51519-042f-44ea-9325-a479be9790d3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1499510217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "909c86730f554a04ab05dcad3a53e9dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28c67e6e-12", "ovs_interfaceid": "28c67e6e-1201-4072-8c7d-331a803bada0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.828614] env[61728]: DEBUG oslo_vmware.api [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464926, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.889475] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.889681] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquired lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.889860] env[61728]: DEBUG nova.network.neutron [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Forcefully refreshing network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1108.976932] env[61728]: DEBUG oslo_vmware.api [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464924, 'name': PowerOffVM_Task, 'duration_secs': 0.1953} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.976932] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1108.976932] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1108.976932] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7365f092-2cb2-499e-b953-e3017c3cad1b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.025020] env[61728]: DEBUG oslo_vmware.api [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464925, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152673} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.025020] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1109.025020] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1109.025254] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1109.041241] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1109.041480] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1109.041673] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleting the datastore file [datastore1] 41329d0d-3838-4d1b-a4e6-8df4508eea76 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1109.042045] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6823b7c6-9ea5-42ad-85e4-8c0a52d60512 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.048239] env[61728]: DEBUG oslo_vmware.api [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1109.048239] env[61728]: value = "task-464928" [ 1109.048239] env[61728]: _type = "Task" [ 1109.048239] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.049188] env[61728]: INFO nova.scheduler.client.report [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleted allocations for instance 77b8f509-17e7-4c25-a43d-1655c16ce8d1 [ 1109.060091] env[61728]: DEBUG oslo_vmware.api [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464928, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.108044] env[61728]: INFO nova.compute.resource_tracker [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating resource usage from migration 915164ff-8e6f-4c15-862f-656699649a53 [ 1109.235892] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36429f72-4346-4a30-bec8-329cd5a0e157 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.243793] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53286011-dae9-47f6-8046-a048b62474a3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.274504] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-346d86de-609c-4e34-98f7-24b7988693ab {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.281855] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b82bf3ee-58fb-4f1d-b84f-8035639adf70 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.301713] env[61728]: DEBUG nova.compute.provider_tree [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1109.325439] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Releasing lock "refresh_cache-79782d2b-1b90-4f3f-a1ce-394e88a00546" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.325859] env[61728]: DEBUG nova.compute.manager [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Instance network_info: |[{"id": "28c67e6e-1201-4072-8c7d-331a803bada0", "address": "fa:16:3e:32:c1:e0", "network": {"id": "4bb51519-042f-44ea-9325-a479be9790d3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1499510217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "909c86730f554a04ab05dcad3a53e9dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28c67e6e-12", "ovs_interfaceid": "28c67e6e-1201-4072-8c7d-331a803bada0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1109.330719] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:c1:e0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '257e5ea7-8b80-4301-9900-a754f1fe2031', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '28c67e6e-1201-4072-8c7d-331a803bada0', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1109.337865] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Creating folder: Project (909c86730f554a04ab05dcad3a53e9dd). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1109.338424] env[61728]: DEBUG oslo_vmware.api [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464926, 'name': PowerOffVM_Task, 'duration_secs': 0.277288} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.338626] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-693c7a19-20b1-4ede-9022-baa06bca2971 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.340265] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1109.340445] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1109.340681] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab7216df-2a2d-4400-90b4-e95191183f4f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.352827] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Created folder: Project (909c86730f554a04ab05dcad3a53e9dd) in parent group-v121913. [ 1109.353024] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Creating folder: Instances. Parent ref: group-v122218. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1109.353249] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb8acaae-337b-4914-bfb3-24652d72b5f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.363459] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Created folder: Instances in parent group-v122218. [ 1109.363698] env[61728]: DEBUG oslo.service.loopingcall [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1109.363891] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1109.364153] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3157988b-6055-4e22-8579-fe352f49840c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.382459] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1109.382459] env[61728]: value = "task-464932" [ 1109.382459] env[61728]: _type = "Task" [ 1109.382459] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.390057] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464932, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.413094] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1109.414059] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1109.414059] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleting the datastore file [datastore1] 40ef29df-12d6-4274-846a-4a2c32752b8b {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1109.414059] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8a4b7172-ceee-4fe3-8ee2-88d800d47458 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.420197] env[61728]: DEBUG oslo_vmware.api [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for the task: (returnval){ [ 1109.420197] env[61728]: value = "task-464933" [ 1109.420197] env[61728]: _type = "Task" [ 1109.420197] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.429764] env[61728]: DEBUG oslo_vmware.api [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464933, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.553620] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.562426] env[61728]: DEBUG oslo_vmware.api [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464928, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13291} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.563023] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1109.563220] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1109.563406] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1109.563583] env[61728]: INFO nova.compute.manager [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1109.563830] env[61728]: DEBUG oslo.service.loopingcall [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1109.564037] env[61728]: DEBUG nova.compute.manager [-] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1109.564134] env[61728]: DEBUG nova.network.neutron [-] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1109.659572] env[61728]: DEBUG nova.compute.manager [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Received event network-changed-28c67e6e-1201-4072-8c7d-331a803bada0 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1109.659793] env[61728]: DEBUG nova.compute.manager [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Refreshing instance network info cache due to event network-changed-28c67e6e-1201-4072-8c7d-331a803bada0. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1109.660047] env[61728]: DEBUG oslo_concurrency.lockutils [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] Acquiring lock "refresh_cache-79782d2b-1b90-4f3f-a1ce-394e88a00546" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.660205] env[61728]: DEBUG oslo_concurrency.lockutils [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] Acquired lock "refresh_cache-79782d2b-1b90-4f3f-a1ce-394e88a00546" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.660374] env[61728]: DEBUG nova.network.neutron [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Refreshing network info cache for port 28c67e6e-1201-4072-8c7d-331a803bada0 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1109.820829] env[61728]: ERROR nova.scheduler.client.report [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [req-c7153cb0-1acc-43e2-9739-3354a0c36632] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c7153cb0-1acc-43e2-9739-3354a0c36632"}]} [ 1109.837386] env[61728]: DEBUG nova.scheduler.client.report [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1109.850936] env[61728]: DEBUG nova.scheduler.client.report [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1109.851191] env[61728]: DEBUG nova.compute.provider_tree [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1109.863041] env[61728]: DEBUG nova.scheduler.client.report [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1109.888075] env[61728]: DEBUG nova.scheduler.client.report [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1109.895952] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464932, 'name': CreateVM_Task, 'duration_secs': 0.475852} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.896176] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1109.903968] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.904207] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.904548] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1109.907084] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0ebcb8a-a9ab-419d-a7b3-f7b89977c26d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.912497] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1109.912497] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5263b714-37af-65ac-5cdf-300e71ddd625" [ 1109.912497] env[61728]: _type = "Task" [ 1109.912497] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.921336] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5263b714-37af-65ac-5cdf-300e71ddd625, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.934949] env[61728]: DEBUG oslo_vmware.api [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Task: {'id': task-464933, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123495} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.935246] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1109.935612] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1109.935846] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1109.936049] env[61728]: INFO nova.compute.manager [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1109.936303] env[61728]: DEBUG oslo.service.loopingcall [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1109.938704] env[61728]: DEBUG nova.compute.manager [-] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1109.938806] env[61728]: DEBUG nova.network.neutron [-] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1110.053795] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b658b9d-8096-4341-a8fe-0c37220f36e2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.061897] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd57774-3f4e-448d-a7d1-e960f16972cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.097416] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e75c2a-a2d0-479a-8d66-7fc787d656e3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.105741] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494b5b5a-2246-4fbe-8d56-794d02798d70 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.120424] env[61728]: DEBUG nova.compute.provider_tree [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1110.203042] env[61728]: DEBUG nova.network.neutron [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Updating instance_info_cache with network_info: [{"id": "51b6c967-37bc-4156-a23f-91040bb1c4a2", "address": "fa:16:3e:7f:f1:5d", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b6c967-37", "ovs_interfaceid": "51b6c967-37bc-4156-a23f-91040bb1c4a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.377063] env[61728]: DEBUG nova.network.neutron [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Updated VIF entry in instance network info cache for port 28c67e6e-1201-4072-8c7d-331a803bada0. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1110.377565] env[61728]: DEBUG nova.network.neutron [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Updating instance_info_cache with network_info: [{"id": "28c67e6e-1201-4072-8c7d-331a803bada0", "address": "fa:16:3e:32:c1:e0", "network": {"id": "4bb51519-042f-44ea-9325-a479be9790d3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1499510217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "909c86730f554a04ab05dcad3a53e9dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28c67e6e-12", "ovs_interfaceid": "28c67e6e-1201-4072-8c7d-331a803bada0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.423562] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5263b714-37af-65ac-5cdf-300e71ddd625, 'name': SearchDatastore_Task, 'duration_secs': 0.012636} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.423880] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.424161] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1110.424415] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.424571] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.424758] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1110.425051] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2afef6d1-4e68-4725-964d-fca7dd4c645f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.433124] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1110.433307] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1110.433986] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df60067e-a334-4b8f-96ec-2d770f8542c4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.439310] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1110.439310] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526ddda0-6e86-9edc-68b4-78d08f793e3a" [ 1110.439310] env[61728]: _type = "Task" [ 1110.439310] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.447270] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526ddda0-6e86-9edc-68b4-78d08f793e3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.499227] env[61728]: DEBUG nova.network.neutron [-] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.654958] env[61728]: DEBUG nova.scheduler.client.report [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 179 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1110.655286] env[61728]: DEBUG nova.compute.provider_tree [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 179 to 180 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1110.655480] env[61728]: DEBUG nova.compute.provider_tree [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1110.706640] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Releasing lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.706863] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Updated the network info_cache for instance {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1110.707072] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.707241] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.748223] env[61728]: DEBUG nova.network.neutron [-] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.856093] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.856093] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.880433] env[61728]: DEBUG oslo_concurrency.lockutils [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] Releasing lock "refresh_cache-79782d2b-1b90-4f3f-a1ce-394e88a00546" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.880748] env[61728]: DEBUG nova.compute.manager [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Received event network-vif-unplugged-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1110.880959] env[61728]: DEBUG oslo_concurrency.lockutils [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] Acquiring lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.881195] env[61728]: DEBUG oslo_concurrency.lockutils [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.881441] env[61728]: DEBUG oslo_concurrency.lockutils [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.881619] env[61728]: DEBUG nova.compute.manager [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] No waiting events found dispatching network-vif-unplugged-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1110.881801] env[61728]: WARNING nova.compute.manager [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Received unexpected event network-vif-unplugged-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb for instance with vm_state shelved_offloaded and task_state None. [ 1110.882035] env[61728]: DEBUG nova.compute.manager [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Received event network-changed-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1110.882222] env[61728]: DEBUG nova.compute.manager [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Refreshing instance network info cache due to event network-changed-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1110.882416] env[61728]: DEBUG oslo_concurrency.lockutils [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] Acquiring lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.882563] env[61728]: DEBUG oslo_concurrency.lockutils [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] Acquired lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.882819] env[61728]: DEBUG nova.network.neutron [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Refreshing network info cache for port ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1110.948806] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526ddda0-6e86-9edc-68b4-78d08f793e3a, 'name': SearchDatastore_Task, 'duration_secs': 0.008632} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.949563] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e97049f-b487-49e1-a0f2-282544babbd9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.954495] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1110.954495] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f645cf-ae5e-9d6b-1575-75941d916ae9" [ 1110.954495] env[61728]: _type = "Task" [ 1110.954495] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.961517] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f645cf-ae5e-9d6b-1575-75941d916ae9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.002214] env[61728]: INFO nova.compute.manager [-] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Took 1.44 seconds to deallocate network for instance. [ 1111.160227] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.064s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.160432] env[61728]: INFO nova.compute.manager [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Migrating [ 1111.171165] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.423s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.172595] env[61728]: INFO nova.compute.claims [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1111.250888] env[61728]: INFO nova.compute.manager [-] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Took 1.31 seconds to deallocate network for instance. [ 1111.464417] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f645cf-ae5e-9d6b-1575-75941d916ae9, 'name': SearchDatastore_Task, 'duration_secs': 0.008572} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.464775] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.464996] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 79782d2b-1b90-4f3f-a1ce-394e88a00546/79782d2b-1b90-4f3f-a1ce-394e88a00546.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1111.465270] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e9e48435-621e-4c2d-8363-26b9b4babaec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.471646] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1111.471646] env[61728]: value = "task-464934" [ 1111.471646] env[61728]: _type = "Task" [ 1111.471646] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.479352] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464934, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.508798] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.611455] env[61728]: DEBUG nova.network.neutron [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Updated VIF entry in instance network info cache for port ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1111.611911] env[61728]: DEBUG nova.network.neutron [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Updating instance_info_cache with network_info: [{"id": "ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb", "address": "fa:16:3e:36:bc:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": null, "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapea6ae87d-d3", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.658312] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.679625] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.679791] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.679946] env[61728]: DEBUG nova.network.neutron [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1111.708507] env[61728]: DEBUG nova.compute.manager [req-b451db1b-45a9-428b-abbe-ceaa7d460da2 req-a0088326-0e8e-4535-a053-06677b6710f7 service nova] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Received event network-vif-deleted-63b2bf95-c628-4c62-bdbd-051192e13a2f {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1111.708507] env[61728]: DEBUG nova.compute.manager [req-b451db1b-45a9-428b-abbe-ceaa7d460da2 req-a0088326-0e8e-4535-a053-06677b6710f7 service nova] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Received event network-vif-deleted-3f69dcc2-f3d8-4b50-b760-85f6849a6c24 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1111.757607] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.854240] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1111.854444] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61728) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1111.854626] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1111.982337] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464934, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454798} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.982519] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 79782d2b-1b90-4f3f-a1ce-394e88a00546/79782d2b-1b90-4f3f-a1ce-394e88a00546.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1111.982733] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1111.982984] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd5122cf-0050-49e8-b3cb-4dbdbc21400e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.988495] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1111.988495] env[61728]: value = "task-464935" [ 1111.988495] env[61728]: _type = "Task" [ 1111.988495] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.996292] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464935, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.114775] env[61728]: DEBUG oslo_concurrency.lockutils [req-f0ec805b-e9e7-44c1-802b-96a604a0bda0 req-85c7df4d-f2e7-4a3c-b54b-e896eb4d6386 service nova] Releasing lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.315843] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1b2fbe-d1ef-4f1a-be08-21ddc97b99a7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.323711] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109b03ff-e559-408c-a21e-7350d7a394c0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.353823] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f421d3-f58e-4f99-b767-723595625911 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.358867] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.363917] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5a35e5-e296-4392-8058-aa8079a12766 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.376625] env[61728]: DEBUG nova.compute.provider_tree [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1112.391773] env[61728]: DEBUG nova.network.neutron [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance_info_cache with network_info: [{"id": "4660e4da-30e8-4e51-9aea-a9375732537f", "address": "fa:16:3e:e1:e9:0c", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4660e4da-30", "ovs_interfaceid": "4660e4da-30e8-4e51-9aea-a9375732537f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.498742] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464935, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063674} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.498986] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1112.499801] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7268a308-780a-4cd7-b9fa-42febfbf5ae0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.522076] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 79782d2b-1b90-4f3f-a1ce-394e88a00546/79782d2b-1b90-4f3f-a1ce-394e88a00546.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1112.522325] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22b4b2d4-74b9-4c7e-855a-399763041984 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.541093] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1112.541093] env[61728]: value = "task-464936" [ 1112.541093] env[61728]: _type = "Task" [ 1112.541093] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.548314] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464936, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.879924] env[61728]: DEBUG nova.scheduler.client.report [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1112.894061] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.050779] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464936, 'name': ReconfigVM_Task, 'duration_secs': 0.297765} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.050906] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 79782d2b-1b90-4f3f-a1ce-394e88a00546/79782d2b-1b90-4f3f-a1ce-394e88a00546.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1113.051562] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-818b6931-27af-42da-83f6-3ec3607903bf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.057518] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1113.057518] env[61728]: value = "task-464937" [ 1113.057518] env[61728]: _type = "Task" [ 1113.057518] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.065458] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464937, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.385381] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.214s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.385875] env[61728]: DEBUG nova.compute.manager [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1113.388629] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.835s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.388806] env[61728]: DEBUG nova.objects.instance [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'resources' on Instance uuid 77b8f509-17e7-4c25-a43d-1655c16ce8d1 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1113.568031] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464937, 'name': Rename_Task, 'duration_secs': 0.141404} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.568386] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1113.568645] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3c7f3e2-bc79-4eb7-8e62-f3c9efa60ffa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.574531] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1113.574531] env[61728]: value = "task-464938" [ 1113.574531] env[61728]: _type = "Task" [ 1113.574531] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.582165] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464938, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.892537] env[61728]: DEBUG nova.compute.utils [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1113.894936] env[61728]: DEBUG nova.objects.instance [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'numa_topology' on Instance uuid 77b8f509-17e7-4c25-a43d-1655c16ce8d1 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1113.895840] env[61728]: DEBUG nova.compute.manager [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Not allocating networking since 'none' was specified. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1114.085258] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464938, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.397027] env[61728]: DEBUG nova.objects.base [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Object Instance<77b8f509-17e7-4c25-a43d-1655c16ce8d1> lazy-loaded attributes: resources,numa_topology {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1114.399908] env[61728]: DEBUG nova.compute.manager [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1114.407304] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e489c9-4595-4c9f-9c85-6bac58081bb2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.426826] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance '8c5b2380-bd14-451e-9612-2767c2a6adb2' progress to 0 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1114.551186] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4405227d-84bf-4bf0-a67d-fd9895b74bad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.560813] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1aaa82-226e-411a-a97f-7b40dbcb7556 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.594778] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b609c91-29ea-41e3-8c44-a1d0191072a2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.605094] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-794d78ed-167e-49c3-b146-71bb87fbca1b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.608996] env[61728]: DEBUG oslo_vmware.api [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464938, 'name': PowerOnVM_Task, 'duration_secs': 0.996192} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.609295] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1114.609503] env[61728]: INFO nova.compute.manager [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Took 7.26 seconds to spawn the instance on the hypervisor. [ 1114.609687] env[61728]: DEBUG nova.compute.manager [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1114.610792] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54dff725-1c69-45ce-9ae9-928b7f792997 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.621284] env[61728]: DEBUG nova.compute.provider_tree [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.934831] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1114.935203] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f704f8f6-a994-4057-8c58-76bec52bd452 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.943209] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1114.943209] env[61728]: value = "task-464939" [ 1114.943209] env[61728]: _type = "Task" [ 1114.943209] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.951904] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464939, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.124832] env[61728]: DEBUG nova.scheduler.client.report [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1115.136937] env[61728]: INFO nova.compute.manager [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Took 13.30 seconds to build instance. [ 1115.409404] env[61728]: DEBUG nova.compute.manager [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1115.430988] env[61728]: DEBUG nova.virt.hardware [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1115.431272] env[61728]: DEBUG nova.virt.hardware [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1115.431437] env[61728]: DEBUG nova.virt.hardware [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1115.431625] env[61728]: DEBUG nova.virt.hardware [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1115.431830] env[61728]: DEBUG nova.virt.hardware [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1115.431994] env[61728]: DEBUG nova.virt.hardware [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1115.432225] env[61728]: DEBUG nova.virt.hardware [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1115.432391] env[61728]: DEBUG nova.virt.hardware [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1115.432566] env[61728]: DEBUG nova.virt.hardware [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1115.432816] env[61728]: DEBUG nova.virt.hardware [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1115.433214] env[61728]: DEBUG nova.virt.hardware [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1115.434139] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca8c067-0748-4dfa-b200-c7288ec7a25b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.442175] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c8058c8-efde-4e55-b04e-790ea6e97331 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.458145] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Instance VIF info [] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1115.463460] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Creating folder: Project (b1561e18018943f3aa87d22c390bfd16). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1115.466230] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0104bb04-27bc-4d0a-bd4d-cfffe7c46595 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.467631] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464939, 'name': PowerOffVM_Task, 'duration_secs': 0.172377} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.467876] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1115.468074] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance '8c5b2380-bd14-451e-9612-2767c2a6adb2' progress to 17 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1115.479780] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Created folder: Project (b1561e18018943f3aa87d22c390bfd16) in parent group-v121913. [ 1115.479968] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Creating folder: Instances. Parent ref: group-v122221. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1115.480204] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0bc070fe-72c3-4676-9b84-cd4507fd51ea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.489195] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Created folder: Instances in parent group-v122221. [ 1115.489425] env[61728]: DEBUG oslo.service.loopingcall [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1115.489608] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1115.489827] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be677095-4da4-4436-af44-7d153b639445 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.506162] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1115.506162] env[61728]: value = "task-464942" [ 1115.506162] env[61728]: _type = "Task" [ 1115.506162] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.514186] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464942, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.629776] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.241s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.632350] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.124s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.632587] env[61728]: DEBUG nova.objects.instance [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lazy-loading 'resources' on Instance uuid 41329d0d-3838-4d1b-a4e6-8df4508eea76 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.638793] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5a8c6eb3-be61-43df-9866-6b2fa01559e9 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "79782d2b-1b90-4f3f-a1ce-394e88a00546" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.813s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.884141] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "532507fa-996d-45c7-bf70-f09c93be79ed" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.884464] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "532507fa-996d-45c7-bf70-f09c93be79ed" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.884672] env[61728]: INFO nova.compute.manager [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Shelving [ 1115.974518] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1115.974905] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1115.974956] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1115.975191] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1115.975351] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1115.975508] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1115.975720] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1115.975885] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1115.976075] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1115.976251] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1115.976433] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1115.981726] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-569fc82e-13bf-40a4-a543-07aff18d1b80 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.997310] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1115.997310] env[61728]: value = "task-464943" [ 1115.997310] env[61728]: _type = "Task" [ 1115.997310] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.005625] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464943, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.014803] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464942, 'name': CreateVM_Task, 'duration_secs': 0.247561} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.014980] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1116.015456] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.015569] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.015894] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1116.016169] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-814c9688-5b93-48ff-baab-f1ebea42ece2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.020859] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1116.020859] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526496d1-c416-a9ab-452e-46cc4e404af2" [ 1116.020859] env[61728]: _type = "Task" [ 1116.020859] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.031650] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526496d1-c416-a9ab-452e-46cc4e404af2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.142038] env[61728]: DEBUG oslo_concurrency.lockutils [None req-aa922879-a111-41cc-9922-66ee96b9c086 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.793s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.146396] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 24.956s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.146396] env[61728]: INFO nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] During sync_power_state the instance has a pending task (shelving_image_pending_upload). Skip. [ 1116.146396] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.146396] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 4.486s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.146396] env[61728]: INFO nova.compute.manager [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Unshelving [ 1116.307907] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7257c8a3-ffef-4c17-b84c-92c92bee61d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.316721] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f640c9-2c5c-4896-973a-a2915bb4d98a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.347265] env[61728]: INFO nova.compute.manager [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Rescuing [ 1116.347543] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "refresh_cache-79782d2b-1b90-4f3f-a1ce-394e88a00546" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.347704] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquired lock "refresh_cache-79782d2b-1b90-4f3f-a1ce-394e88a00546" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.347880] env[61728]: DEBUG nova.network.neutron [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1116.349891] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1677403-773a-4529-9122-16d8a7d97883 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.357598] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f589458-10bb-4009-bc22-98a04096a5b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.372228] env[61728]: DEBUG nova.compute.provider_tree [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.391823] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1116.392083] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f72e088-46d4-424e-8d81-a361e8776c1f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.399444] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1116.399444] env[61728]: value = "task-464944" [ 1116.399444] env[61728]: _type = "Task" [ 1116.399444] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.407869] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464944, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.508781] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464943, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.533660] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]526496d1-c416-a9ab-452e-46cc4e404af2, 'name': SearchDatastore_Task, 'duration_secs': 0.01046} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.534094] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.534390] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1116.534647] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.534792] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.534987] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1116.535266] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15491090-93f9-432b-9dd8-fb624bafd744 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.543683] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1116.543868] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1116.544592] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9703b6a-2bd6-4cb2-be67-7923286c4440 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.550129] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1116.550129] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527a6f63-4134-e974-9294-2decd98e234d" [ 1116.550129] env[61728]: _type = "Task" [ 1116.550129] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.558297] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527a6f63-4134-e974-9294-2decd98e234d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.875943] env[61728]: DEBUG nova.scheduler.client.report [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1116.909115] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464944, 'name': PowerOffVM_Task, 'duration_secs': 0.233571} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.909607] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1116.910222] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30e8fa1-f20b-410a-9e9b-e58028769d9d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.929631] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9f5328-18fd-44d2-a92d-3a5f05be90fa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.010221] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464943, 'name': ReconfigVM_Task, 'duration_secs': 0.87482} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.010561] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance '8c5b2380-bd14-451e-9612-2767c2a6adb2' progress to 33 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1117.056557] env[61728]: DEBUG nova.network.neutron [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Updating instance_info_cache with network_info: [{"id": "28c67e6e-1201-4072-8c7d-331a803bada0", "address": "fa:16:3e:32:c1:e0", "network": {"id": "4bb51519-042f-44ea-9325-a479be9790d3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1499510217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "909c86730f554a04ab05dcad3a53e9dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28c67e6e-12", "ovs_interfaceid": "28c67e6e-1201-4072-8c7d-331a803bada0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.062030] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527a6f63-4134-e974-9294-2decd98e234d, 'name': SearchDatastore_Task, 'duration_secs': 0.008788} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.063089] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d1f525b-040c-47b9-b6f9-2f236645f35b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.068356] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1117.068356] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a31f79-7448-9b0a-641b-dd64edf95885" [ 1117.068356] env[61728]: _type = "Task" [ 1117.068356] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.076959] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a31f79-7448-9b0a-641b-dd64edf95885, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.166874] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.380877] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.748s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.383774] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.626s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.384065] env[61728]: DEBUG nova.objects.instance [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lazy-loading 'resources' on Instance uuid 40ef29df-12d6-4274-846a-4a2c32752b8b {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.398723] env[61728]: INFO nova.scheduler.client.report [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleted allocations for instance 41329d0d-3838-4d1b-a4e6-8df4508eea76 [ 1117.439544] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1117.439901] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9e244d04-8e4f-4121-a668-25289c008d96 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.448721] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1117.448721] env[61728]: value = "task-464945" [ 1117.448721] env[61728]: _type = "Task" [ 1117.448721] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.458878] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464945, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.517988] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1117.518317] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1117.518949] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1117.518949] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1117.518949] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1117.519265] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1117.519320] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1117.519456] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1117.519664] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1117.519951] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1117.520250] env[61728]: DEBUG nova.virt.hardware [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1117.525626] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Reconfiguring VM instance instance-00000058 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1117.526283] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0425fb51-91ec-4f36-81c5-018ad87d0b0f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.544151] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1117.544151] env[61728]: value = "task-464946" [ 1117.544151] env[61728]: _type = "Task" [ 1117.544151] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.552201] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464946, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.563822] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Releasing lock "refresh_cache-79782d2b-1b90-4f3f-a1ce-394e88a00546" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.578535] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a31f79-7448-9b0a-641b-dd64edf95885, 'name': SearchDatastore_Task, 'duration_secs': 0.008568} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.578801] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.579094] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] dbe877e6-1754-4cf7-b454-abd88bfa713a/dbe877e6-1754-4cf7-b454-abd88bfa713a.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1117.579364] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad050359-8df2-4ad4-b085-81b3cf25e688 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.586461] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1117.586461] env[61728]: value = "task-464947" [ 1117.586461] env[61728]: _type = "Task" [ 1117.586461] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.597639] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464947, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.906477] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ad684a81-8390-42d1-aa24-47654bd45ebe tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "41329d0d-3838-4d1b-a4e6-8df4508eea76" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.459s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.961385] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464945, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.051257] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419d0a87-deec-4ee8-b811-8571cb86dbcb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.062435] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfc9966-c354-4b6f-a6ca-3b451495e938 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.065724] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464946, 'name': ReconfigVM_Task, 'duration_secs': 0.169011} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.066027] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Reconfigured VM instance instance-00000058 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1118.067179] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0b06bc-3066-4efc-bd76-7d1d96e4c295 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.107174] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a70b6a-6503-45a2-91ba-ea3208a23675 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.109935] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1118.122815] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1eee5878-d197-4b3c-bd15-ad4d42cfc3fe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.131959] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 8c5b2380-bd14-451e-9612-2767c2a6adb2/8c5b2380-bd14-451e-9612-2767c2a6adb2.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1118.132949] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-922573b5-29aa-44ac-af00-a4a5d44ec9b4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.151945] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464947, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.4457} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.153887] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79eaa25-90f7-4d5a-94f8-c2201bcb5936 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.158271] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] dbe877e6-1754-4cf7-b454-abd88bfa713a/dbe877e6-1754-4cf7-b454-abd88bfa713a.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1118.158509] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1118.160647] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8caf82f4-07a4-4267-b291-0248e9b6cdcf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.162703] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1118.162703] env[61728]: value = "task-464949" [ 1118.162703] env[61728]: _type = "Task" [ 1118.162703] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.163085] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1118.163085] env[61728]: value = "task-464948" [ 1118.163085] env[61728]: _type = "Task" [ 1118.163085] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.181046] env[61728]: DEBUG nova.compute.provider_tree [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1118.182958] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1118.182958] env[61728]: value = "task-464950" [ 1118.182958] env[61728]: _type = "Task" [ 1118.182958] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.194991] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464948, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.195376] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464949, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.200863] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464950, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.461145] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464945, 'name': CreateSnapshot_Task, 'duration_secs': 0.667293} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.461480] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1118.462250] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-288bb203-433b-4302-8389-9cb022e73f31 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.676645] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464948, 'name': PowerOffVM_Task, 'duration_secs': 0.196333} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.679827] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1118.679827] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464949, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.680505] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da44fdd-b41e-4bb0-9af1-454eca0a97f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.684768] env[61728]: DEBUG nova.scheduler.client.report [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1118.703176] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.319s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.708018] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0467ae6e-e97c-4423-bd39-82d7b4b87364 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.710564] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.352s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.710741] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.710904] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61728) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1118.711170] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.545s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.711376] env[61728]: DEBUG nova.objects.instance [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'pci_requests' on Instance uuid 77b8f509-17e7-4c25-a43d-1655c16ce8d1 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1118.713137] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a0beab-b3bf-4a33-bfc3-1fe408c24c97 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.720292] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464950, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080745} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.724509] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1118.727864] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9906ab-16f1-4a0b-a7aa-9c6cb3827476 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.730811] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e03039b-a524-48cb-9096-d7571676434f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.735334] env[61728]: INFO nova.scheduler.client.report [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Deleted allocations for instance 40ef29df-12d6-4274-846a-4a2c32752b8b [ 1118.751288] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1118.758604] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6fe5aa-5ddb-4913-a93a-042cabebef91 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.761093] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c20252aa-525a-43f1-ab3d-0e4a95977d4e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.771165] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] dbe877e6-1754-4cf7-b454-abd88bfa713a/dbe877e6-1754-4cf7-b454-abd88bfa713a.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1118.771450] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a7d7213-6dda-4991-9455-aa61bfc1e8e9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.790780] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5594bb3-21d8-4ea4-baab-d0b72b4e31fb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.796090] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1118.796090] env[61728]: value = "task-464952" [ 1118.796090] env[61728]: _type = "Task" [ 1118.796090] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.796351] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1118.796351] env[61728]: value = "task-464951" [ 1118.796351] env[61728]: _type = "Task" [ 1118.796351] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.826378] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180570MB free_disk=115GB free_vcpus=48 pci_devices=None {{(pid=61728) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1118.826608] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.831759] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464952, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.835468] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1118.835700] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1118.835949] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.836123] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.836311] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1118.836563] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7bb2bb02-ce73-4158-a03f-89c00d6f08a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.844675] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1118.844866] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1118.845571] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c58eb2d0-014d-4c46-ab51-b05cba89525b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.850713] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1118.850713] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5263eb89-aacc-14e1-8a23-d45fd158ae3e" [ 1118.850713] env[61728]: _type = "Task" [ 1118.850713] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.858111] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5263eb89-aacc-14e1-8a23-d45fd158ae3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.980488] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1118.980833] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c10a29fb-d57a-4fff-9aed-ed6d1c4cb79a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.989706] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1118.989706] env[61728]: value = "task-464953" [ 1118.989706] env[61728]: _type = "Task" [ 1118.989706] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.997755] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464953, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.175576] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464949, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.217908] env[61728]: DEBUG nova.objects.instance [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'numa_topology' on Instance uuid 77b8f509-17e7-4c25-a43d-1655c16ce8d1 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.244904] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b2e5af93-3f58-42ae-a76e-da90710b675a tempest-ServersTestJSON-1395267504 tempest-ServersTestJSON-1395267504-project-member] Lock "40ef29df-12d6-4274-846a-4a2c32752b8b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.445s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.307424] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464952, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.361473] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5263eb89-aacc-14e1-8a23-d45fd158ae3e, 'name': SearchDatastore_Task, 'duration_secs': 0.009584} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.362325] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d119062c-7cac-4bd8-ae19-d2e3f255084e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.368159] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1119.368159] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5222249b-456a-c3c6-67da-22a6edf88ed2" [ 1119.368159] env[61728]: _type = "Task" [ 1119.368159] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.376932] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5222249b-456a-c3c6-67da-22a6edf88ed2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.501165] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464953, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.600268] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "2db96467-5fb4-49f4-954b-dbdd21e46232" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.600493] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "2db96467-5fb4-49f4-954b-dbdd21e46232" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.678232] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464949, 'name': ReconfigVM_Task, 'duration_secs': 1.319707} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.678616] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 8c5b2380-bd14-451e-9612-2767c2a6adb2/8c5b2380-bd14-451e-9612-2767c2a6adb2.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1119.678983] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance '8c5b2380-bd14-451e-9612-2767c2a6adb2' progress to 50 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1119.720324] env[61728]: INFO nova.compute.claims [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1119.809605] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464952, 'name': ReconfigVM_Task, 'duration_secs': 0.907239} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.810033] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Reconfigured VM instance instance-00000075 to attach disk [datastore1] dbe877e6-1754-4cf7-b454-abd88bfa713a/dbe877e6-1754-4cf7-b454-abd88bfa713a.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1119.810738] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-727ce593-4bd0-4ee4-95c8-07cb16db4edf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.817557] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1119.817557] env[61728]: value = "task-464954" [ 1119.817557] env[61728]: _type = "Task" [ 1119.817557] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.830800] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464954, 'name': Rename_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.878796] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5222249b-456a-c3c6-67da-22a6edf88ed2, 'name': SearchDatastore_Task, 'duration_secs': 0.008823} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.879115] env[61728]: DEBUG oslo_concurrency.lockutils [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.879455] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 79782d2b-1b90-4f3f-a1ce-394e88a00546/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk. {{(pid=61728) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1119.879788] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d073736-d200-4fcf-9b75-de5e00391f90 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.886742] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1119.886742] env[61728]: value = "task-464955" [ 1119.886742] env[61728]: _type = "Task" [ 1119.886742] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.895296] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464955, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.003054] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464953, 'name': CloneVM_Task} progress is 95%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.103234] env[61728]: DEBUG nova.compute.manager [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1120.187696] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80660056-5d74-4094-8966-1bfd04e6420b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.208936] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b5bc38-a2c9-46d5-b2a2-8f0be6ddd02e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.233185] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance '8c5b2380-bd14-451e-9612-2767c2a6adb2' progress to 67 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1120.331868] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464954, 'name': Rename_Task, 'duration_secs': 0.162604} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.332249] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1120.332521] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c06d0958-f912-4489-9a9b-60db0164eda2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.340787] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1120.340787] env[61728]: value = "task-464956" [ 1120.340787] env[61728]: _type = "Task" [ 1120.340787] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.354339] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464956, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.396227] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464955, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.502398] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464953, 'name': CloneVM_Task, 'duration_secs': 1.193978} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.502699] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Created linked-clone VM from snapshot [ 1120.503501] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b699a30-c34a-4a8c-9a2c-9b15dab7351b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.511087] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Uploading image df2f6391-a88d-4bb9-ada0-dd01a6acec62 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1120.539982] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1120.539982] env[61728]: value = "vm-122225" [ 1120.539982] env[61728]: _type = "VirtualMachine" [ 1120.539982] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1120.540320] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ca8a3e4c-7245-44ca-bd8a-a07c050a0e68 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.547417] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lease: (returnval){ [ 1120.547417] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52739b4f-3dd1-6782-72ff-db7535527794" [ 1120.547417] env[61728]: _type = "HttpNfcLease" [ 1120.547417] env[61728]: } obtained for exporting VM: (result){ [ 1120.547417] env[61728]: value = "vm-122225" [ 1120.547417] env[61728]: _type = "VirtualMachine" [ 1120.547417] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1120.547663] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the lease: (returnval){ [ 1120.547663] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52739b4f-3dd1-6782-72ff-db7535527794" [ 1120.547663] env[61728]: _type = "HttpNfcLease" [ 1120.547663] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1120.554319] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1120.554319] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52739b4f-3dd1-6782-72ff-db7535527794" [ 1120.554319] env[61728]: _type = "HttpNfcLease" [ 1120.554319] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1120.627930] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.799858] env[61728]: DEBUG nova.network.neutron [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Port 4660e4da-30e8-4e51-9aea-a9375732537f binding to destination host cpu-1 is already ACTIVE {{(pid=61728) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1120.853897] env[61728]: DEBUG oslo_vmware.api [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464956, 'name': PowerOnVM_Task, 'duration_secs': 0.43459} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.854249] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1120.854487] env[61728]: INFO nova.compute.manager [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Took 5.44 seconds to spawn the instance on the hypervisor. [ 1120.854834] env[61728]: DEBUG nova.compute.manager [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1120.855635] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3edd725-25b6-4b2f-b125-ca3d4b93981e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.893326] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c85ac27-1347-492a-9f26-60b3109ad57d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.900822] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464955, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.903812] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c88df71-7bbf-4872-b705-871f88b4b9d2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.934583] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51f769d-2fc0-40a5-a1cc-8f001488d858 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.942087] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3975697-098d-4e89-8779-fbffbcba25f2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.958041] env[61728]: DEBUG nova.compute.provider_tree [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1121.056507] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1121.056507] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52739b4f-3dd1-6782-72ff-db7535527794" [ 1121.056507] env[61728]: _type = "HttpNfcLease" [ 1121.056507] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1121.056838] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1121.056838] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52739b4f-3dd1-6782-72ff-db7535527794" [ 1121.056838] env[61728]: _type = "HttpNfcLease" [ 1121.056838] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1121.057660] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0496e7fe-e94d-4cf6-93bf-3aa8edc7c5c8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.065439] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e413d2-fd60-5852-0a38-b7449bd5f849/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1121.065624] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e413d2-fd60-5852-0a38-b7449bd5f849/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1121.191012] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e84d81b8-bb9b-4fc2-b895-b42bf688d3bf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.376423] env[61728]: INFO nova.compute.manager [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Took 13.64 seconds to build instance. [ 1121.399506] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464955, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.460699] env[61728]: DEBUG nova.scheduler.client.report [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1121.795295] env[61728]: INFO nova.compute.manager [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Rebuilding instance [ 1121.824829] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "8c5b2380-bd14-451e-9612-2767c2a6adb2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.825135] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.825703] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.856318] env[61728]: DEBUG nova.compute.manager [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1121.857577] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9996d1de-6ae3-45e9-9688-127f000eed6e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.878469] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b5af8f42-b7e3-4398-b299-bdd4356d7347 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Lock "dbe877e6-1754-4cf7-b454-abd88bfa713a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.151s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.898456] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464955, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.513097} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.899548] env[61728]: INFO nova.virt.vmwareapi.ds_util [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 79782d2b-1b90-4f3f-a1ce-394e88a00546/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk. [ 1121.900739] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50a4e1c-a490-4774-a69a-40a1b44267db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.933704] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 79782d2b-1b90-4f3f-a1ce-394e88a00546/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1121.934616] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e5373a5-82ee-4d39-b2c9-d36543b563ce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.954634] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1121.954634] env[61728]: value = "task-464958" [ 1121.954634] env[61728]: _type = "Task" [ 1121.954634] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.966322] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.255s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.968789] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464958, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.969151] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.143s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.999597] env[61728]: INFO nova.network.neutron [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Updating port ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1122.369090] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1122.369480] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad3e03c1-b334-4a41-822a-4adc56f722cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.379106] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1122.379106] env[61728]: value = "task-464959" [ 1122.379106] env[61728]: _type = "Task" [ 1122.379106] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.387457] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464959, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.466264] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464958, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.858863] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.859159] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.859410] env[61728]: DEBUG nova.network.neutron [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1122.889622] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464959, 'name': PowerOffVM_Task, 'duration_secs': 0.192481} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.889995] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1122.890304] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1122.891198] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97bf27c3-5530-4b2a-8abe-ee1780b6826b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.897945] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1122.898233] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b8578c8-b8bf-4393-bc17-8f88356756cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.921959] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1122.922326] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1122.922597] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Deleting the datastore file [datastore1] dbe877e6-1754-4cf7-b454-abd88bfa713a {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1122.922873] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c4eb4b7-0c4a-45d0-b7e9-01751932ee42 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.929605] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1122.929605] env[61728]: value = "task-464961" [ 1122.929605] env[61728]: _type = "Task" [ 1122.929605] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.937814] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464961, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.965687] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464958, 'name': ReconfigVM_Task, 'duration_secs': 0.584517} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.966124] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 79782d2b-1b90-4f3f-a1ce-394e88a00546/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1122.967108] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9680a99d-b1d9-4ac7-80f8-d0064165f3a9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.988473] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Applying migration context for instance 8c5b2380-bd14-451e-9612-2767c2a6adb2 as it has an incoming, in-progress migration 915164ff-8e6f-4c15-862f-656699649a53. Migration status is post-migrating {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1122.989944] env[61728]: INFO nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating resource usage from migration 915164ff-8e6f-4c15-862f-656699649a53 [ 1122.997895] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8eabecb8-7822-42ce-b805-cd2e5ca5714f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.015936] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1123.015936] env[61728]: value = "task-464962" [ 1123.015936] env[61728]: _type = "Task" [ 1123.015936] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.017391] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance b1fb2e3f-234a-44c7-983e-d4441b4e3cef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1123.018117] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance d55beb02-00fc-4df6-a239-9e5d776a7c18 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1123.018117] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 532507fa-996d-45c7-bf70-f09c93be79ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1123.018117] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance c1ed0e8b-5cf9-4ff0-931e-fadc3432a125 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1123.018117] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 79782d2b-1b90-4f3f-a1ce-394e88a00546 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1123.018492] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Migration 915164ff-8e6f-4c15-862f-656699649a53 is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1123.018902] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 8c5b2380-bd14-451e-9612-2767c2a6adb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1123.019176] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance dbe877e6-1754-4cf7-b454-abd88bfa713a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1123.020224] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 77b8f509-17e7-4c25-a43d-1655c16ce8d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1123.031339] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464962, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.431151] env[61728]: DEBUG nova.compute.manager [req-4535adea-c7b3-4267-a5e2-b7ddec7118d2 req-eec7d6e6-d0ab-44e8-8819-b5d07fca4255 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Received event network-vif-plugged-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1123.431745] env[61728]: DEBUG oslo_concurrency.lockutils [req-4535adea-c7b3-4267-a5e2-b7ddec7118d2 req-eec7d6e6-d0ab-44e8-8819-b5d07fca4255 service nova] Acquiring lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.431745] env[61728]: DEBUG oslo_concurrency.lockutils [req-4535adea-c7b3-4267-a5e2-b7ddec7118d2 req-eec7d6e6-d0ab-44e8-8819-b5d07fca4255 service nova] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.432074] env[61728]: DEBUG oslo_concurrency.lockutils [req-4535adea-c7b3-4267-a5e2-b7ddec7118d2 req-eec7d6e6-d0ab-44e8-8819-b5d07fca4255 service nova] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.432288] env[61728]: DEBUG nova.compute.manager [req-4535adea-c7b3-4267-a5e2-b7ddec7118d2 req-eec7d6e6-d0ab-44e8-8819-b5d07fca4255 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] No waiting events found dispatching network-vif-plugged-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1123.432415] env[61728]: WARNING nova.compute.manager [req-4535adea-c7b3-4267-a5e2-b7ddec7118d2 req-eec7d6e6-d0ab-44e8-8819-b5d07fca4255 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Received unexpected event network-vif-plugged-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb for instance with vm_state shelved_offloaded and task_state spawning. [ 1123.443879] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464961, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.087094} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.444195] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1123.444469] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1123.444775] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1123.527199] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 2db96467-5fb4-49f4-954b-dbdd21e46232 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1123.527469] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1123.527638] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2304MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1123.539716] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464962, 'name': ReconfigVM_Task, 'duration_secs': 0.230567} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.541718] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1123.542048] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff4b2641-0d92-417a-b686-ff4ac16c1f0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.550579] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1123.550579] env[61728]: value = "task-464963" [ 1123.550579] env[61728]: _type = "Task" [ 1123.550579] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.562208] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464963, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.600871] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.602572] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.602572] env[61728]: DEBUG nova.network.neutron [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1123.679290] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7b4ada-b090-41f4-998c-8b3f63eefeed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.684580] env[61728]: DEBUG nova.network.neutron [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance_info_cache with network_info: [{"id": "4660e4da-30e8-4e51-9aea-a9375732537f", "address": "fa:16:3e:e1:e9:0c", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4660e4da-30", "ovs_interfaceid": "4660e4da-30e8-4e51-9aea-a9375732537f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.689073] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e780f5d-a984-4f0c-8bd4-262548ab7f2b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.721694] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab3c2a3-54c1-4790-8773-1aee41279351 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.729264] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df0aaa2f-0158-4eff-a86c-b50953768710 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.745060] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1124.063136] env[61728]: DEBUG oslo_vmware.api [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-464963, 'name': PowerOnVM_Task, 'duration_secs': 0.484441} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.063458] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1124.066414] env[61728]: DEBUG nova.compute.manager [None req-10b262bb-c585-4127-a9d0-1f6c51f1f607 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1124.067267] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a75ef54a-6548-4a27-9522-656e889cb4d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.187435] env[61728]: DEBUG oslo_concurrency.lockutils [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.247467] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1124.333016] env[61728]: DEBUG nova.network.neutron [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Updating instance_info_cache with network_info: [{"id": "ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb", "address": "fa:16:3e:36:bc:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea6ae87d-d3", "ovs_interfaceid": "ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.481960] env[61728]: DEBUG nova.virt.hardware [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1124.482308] env[61728]: DEBUG nova.virt.hardware [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1124.482416] env[61728]: DEBUG nova.virt.hardware [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1124.482816] env[61728]: DEBUG nova.virt.hardware [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1124.482816] env[61728]: DEBUG nova.virt.hardware [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1124.482964] env[61728]: DEBUG nova.virt.hardware [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1124.483305] env[61728]: DEBUG nova.virt.hardware [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1124.483506] env[61728]: DEBUG nova.virt.hardware [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1124.483690] env[61728]: DEBUG nova.virt.hardware [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1124.483864] env[61728]: DEBUG nova.virt.hardware [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1124.484059] env[61728]: DEBUG nova.virt.hardware [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1124.485092] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca02436-1af6-4da6-be3f-3cd4465b61ea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.494142] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081504cf-f7fe-4636-bae4-b03969288d25 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.508106] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Instance VIF info [] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1124.514202] env[61728]: DEBUG oslo.service.loopingcall [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1124.514836] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1124.514836] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-380eb441-7ec5-426e-a111-a00439bb023a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.532729] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1124.532729] env[61728]: value = "task-464964" [ 1124.532729] env[61728]: _type = "Task" [ 1124.532729] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.540889] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464964, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.711929] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e4d2664-59a0-4d09-8f9c-69f2f60b3815 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.734192] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e53cafe7-3b23-4ad1-97d0-542db3632023 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.742553] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance '8c5b2380-bd14-451e-9612-2767c2a6adb2' progress to 83 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1124.752573] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1124.752826] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.784s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.753161] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.125s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.754705] env[61728]: INFO nova.compute.claims [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1124.836252] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.859663] env[61728]: DEBUG nova.virt.hardware [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='5fa368ea6c1bb8ee83198c6c69496cd2',container_format='bare',created_at=2024-08-29T12:32:27Z,direct_url=,disk_format='vmdk',id=d9914484-11b1-4819-982a-67a108bd3ed1,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1629180430-shelved',owner='49c64edaa37f41d38aba7bd5b1d0b47f',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-08-29T12:32:43Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1124.859973] env[61728]: DEBUG nova.virt.hardware [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1124.860181] env[61728]: DEBUG nova.virt.hardware [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1124.860400] env[61728]: DEBUG nova.virt.hardware [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1124.860561] env[61728]: DEBUG nova.virt.hardware [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1124.860716] env[61728]: DEBUG nova.virt.hardware [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1124.860965] env[61728]: DEBUG nova.virt.hardware [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1124.861200] env[61728]: DEBUG nova.virt.hardware [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1124.861390] env[61728]: DEBUG nova.virt.hardware [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1124.861562] env[61728]: DEBUG nova.virt.hardware [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1124.861769] env[61728]: DEBUG nova.virt.hardware [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1124.862956] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d856aba-5f09-4ca0-b913-6d96400c5474 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.871470] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e12d0271-5535-4471-9c80-8d459c784067 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.885453] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:bc:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e51ebca-e0f8-4b77-b155-4ff928eef130', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1124.892829] env[61728]: DEBUG oslo.service.loopingcall [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1124.893152] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1124.893394] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dcef2eaa-d89c-4e56-908c-1c1fd8838f7a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.912293] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1124.912293] env[61728]: value = "task-464965" [ 1124.912293] env[61728]: _type = "Task" [ 1124.912293] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.921098] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464965, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.044746] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464964, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.249156] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1125.249677] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9e7af191-4f68-4ccd-89a9-7bf039ff5786 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.256959] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1125.256959] env[61728]: value = "task-464966" [ 1125.256959] env[61728]: _type = "Task" [ 1125.256959] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.267138] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464966, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.422418] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464965, 'name': CreateVM_Task, 'duration_secs': 0.333123} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.422692] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1125.423365] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d9914484-11b1-4819-982a-67a108bd3ed1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.423545] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d9914484-11b1-4819-982a-67a108bd3ed1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.423918] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d9914484-11b1-4819-982a-67a108bd3ed1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1125.424195] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-326a40a8-19ac-4c74-adb3-8e600d2a3f4d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.430208] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1125.430208] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c237fc-8221-9faa-2bd5-540aa688de20" [ 1125.430208] env[61728]: _type = "Task" [ 1125.430208] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.438730] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c237fc-8221-9faa-2bd5-540aa688de20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.475706] env[61728]: DEBUG nova.compute.manager [req-35bdfeba-cc34-437a-9b7d-b3c5a9cc4141 req-13969392-b49e-413a-b069-7f740d37b461 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Received event network-changed-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1125.476067] env[61728]: DEBUG nova.compute.manager [req-35bdfeba-cc34-437a-9b7d-b3c5a9cc4141 req-13969392-b49e-413a-b069-7f740d37b461 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Refreshing instance network info cache due to event network-changed-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1125.476302] env[61728]: DEBUG oslo_concurrency.lockutils [req-35bdfeba-cc34-437a-9b7d-b3c5a9cc4141 req-13969392-b49e-413a-b069-7f740d37b461 service nova] Acquiring lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.476531] env[61728]: DEBUG oslo_concurrency.lockutils [req-35bdfeba-cc34-437a-9b7d-b3c5a9cc4141 req-13969392-b49e-413a-b069-7f740d37b461 service nova] Acquired lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.476740] env[61728]: DEBUG nova.network.neutron [req-35bdfeba-cc34-437a-9b7d-b3c5a9cc4141 req-13969392-b49e-413a-b069-7f740d37b461 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Refreshing network info cache for port ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1125.543879] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464964, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.768994] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464966, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.903328] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54cc8006-7b64-4dbc-9531-da305f682fe2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.911153] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-346b91f3-fd40-45cc-a833-76322b100a0f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.945584] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-268e1ca6-323e-42c3-87e0-049b2e28e9e0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.956339] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d9914484-11b1-4819-982a-67a108bd3ed1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.956655] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Processing image d9914484-11b1-4819-982a-67a108bd3ed1 {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1125.958975] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d9914484-11b1-4819-982a-67a108bd3ed1/d9914484-11b1-4819-982a-67a108bd3ed1.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.958975] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d9914484-11b1-4819-982a-67a108bd3ed1/d9914484-11b1-4819-982a-67a108bd3ed1.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.958975] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1125.958975] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70230202-a51c-408a-8a6f-43a87dc666e3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.960397] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef14e2a-bb6b-42ef-90ba-3b50ffc9b1b5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.975685] env[61728]: DEBUG nova.compute.provider_tree [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.977990] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1125.978159] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1125.981062] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e410df7-2e35-4a64-8c7c-83fae79de7d0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.987255] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1125.987255] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a7b0d2-955a-bd5b-3f41-bbcd647a1382" [ 1125.987255] env[61728]: _type = "Task" [ 1125.987255] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.995697] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52a7b0d2-955a-bd5b-3f41-bbcd647a1382, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.019521] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "c92e595d-c686-4810-bb39-d98de4141121" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.019785] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "c92e595d-c686-4810-bb39-d98de4141121" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.045328] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464964, 'name': CreateVM_Task, 'duration_secs': 1.31378} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.045512] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1126.045950] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.046137] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.046484] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1126.046753] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99faf1a9-8a68-4ec1-8a56-ebb1d1520a54 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.054233] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1126.054233] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aef22e-5beb-11ba-0ee2-5acc29a5a15e" [ 1126.054233] env[61728]: _type = "Task" [ 1126.054233] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.062567] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aef22e-5beb-11ba-0ee2-5acc29a5a15e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.235818] env[61728]: DEBUG nova.network.neutron [req-35bdfeba-cc34-437a-9b7d-b3c5a9cc4141 req-13969392-b49e-413a-b069-7f740d37b461 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Updated VIF entry in instance network info cache for port ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1126.236259] env[61728]: DEBUG nova.network.neutron [req-35bdfeba-cc34-437a-9b7d-b3c5a9cc4141 req-13969392-b49e-413a-b069-7f740d37b461 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Updating instance_info_cache with network_info: [{"id": "ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb", "address": "fa:16:3e:36:bc:84", "network": {"id": "555e33ff-b003-4786-8f62-9fc4f52e2682", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-712876823-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c64edaa37f41d38aba7bd5b1d0b47f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea6ae87d-d3", "ovs_interfaceid": "ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.268795] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464966, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.480610] env[61728]: DEBUG nova.scheduler.client.report [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1126.499577] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Preparing fetch location {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1126.499753] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Fetch image to [datastore1] OSTACK_IMG_e1bbc824-f565-4bca-8a9a-f6cb7551993c/OSTACK_IMG_e1bbc824-f565-4bca-8a9a-f6cb7551993c.vmdk {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1126.499939] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Downloading stream optimized image d9914484-11b1-4819-982a-67a108bd3ed1 to [datastore1] OSTACK_IMG_e1bbc824-f565-4bca-8a9a-f6cb7551993c/OSTACK_IMG_e1bbc824-f565-4bca-8a9a-f6cb7551993c.vmdk on the data store datastore1 as vApp {{(pid=61728) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1126.500157] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Downloading image file data d9914484-11b1-4819-982a-67a108bd3ed1 to the ESX as VM named 'OSTACK_IMG_e1bbc824-f565-4bca-8a9a-f6cb7551993c' {{(pid=61728) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1126.523906] env[61728]: DEBUG nova.compute.manager [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1126.565218] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52aef22e-5beb-11ba-0ee2-5acc29a5a15e, 'name': SearchDatastore_Task, 'duration_secs': 0.018123} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.567574] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.567857] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1126.568118] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.568281] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.568472] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1126.570729] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a8a5804e-a034-4ef5-8475-2241fc1fc209 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.577896] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1126.578122] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1126.579155] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1df9ee04-af85-46ca-b630-28282d3748f7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.584808] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1126.584808] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52baa6fc-e797-d7e4-6dca-c134c3a55396" [ 1126.584808] env[61728]: _type = "Task" [ 1126.584808] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.593428] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52baa6fc-e797-d7e4-6dca-c134c3a55396, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.594341] env[61728]: DEBUG oslo_vmware.rw_handles [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1126.594341] env[61728]: value = "resgroup-9" [ 1126.594341] env[61728]: _type = "ResourcePool" [ 1126.594341] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1126.594595] env[61728]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-2d8e3d11-6edc-4d6a-8d2c-dd60335ab4f8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.614693] env[61728]: DEBUG oslo_vmware.rw_handles [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lease: (returnval){ [ 1126.614693] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52af5b56-eb34-eb0d-eb90-e35d955cff28" [ 1126.614693] env[61728]: _type = "HttpNfcLease" [ 1126.614693] env[61728]: } obtained for vApp import into resource pool (val){ [ 1126.614693] env[61728]: value = "resgroup-9" [ 1126.614693] env[61728]: _type = "ResourcePool" [ 1126.614693] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1126.615071] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the lease: (returnval){ [ 1126.615071] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52af5b56-eb34-eb0d-eb90-e35d955cff28" [ 1126.615071] env[61728]: _type = "HttpNfcLease" [ 1126.615071] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1126.621080] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1126.621080] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52af5b56-eb34-eb0d-eb90-e35d955cff28" [ 1126.621080] env[61728]: _type = "HttpNfcLease" [ 1126.621080] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1126.739034] env[61728]: DEBUG oslo_concurrency.lockutils [req-35bdfeba-cc34-437a-9b7d-b3c5a9cc4141 req-13969392-b49e-413a-b069-7f740d37b461 service nova] Releasing lock "refresh_cache-77b8f509-17e7-4c25-a43d-1655c16ce8d1" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.768060] env[61728]: DEBUG oslo_vmware.api [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464966, 'name': PowerOnVM_Task, 'duration_secs': 1.037366} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.768368] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1126.768569] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-eeb6468b-23a5-4271-9e9f-ac959eae4379 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance '8c5b2380-bd14-451e-9612-2767c2a6adb2' progress to 100 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1126.986392] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.233s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.986979] env[61728]: DEBUG nova.compute.manager [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1127.047951] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.048326] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.050402] env[61728]: INFO nova.compute.claims [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1127.096252] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52baa6fc-e797-d7e4-6dca-c134c3a55396, 'name': SearchDatastore_Task, 'duration_secs': 0.00894} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.097123] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7192f7cf-faec-4eeb-a050-15958b6a02dd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.102687] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1127.102687] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52beeba9-501b-3c3f-d1e7-3c7629dc5072" [ 1127.102687] env[61728]: _type = "Task" [ 1127.102687] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.112401] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52beeba9-501b-3c3f-d1e7-3c7629dc5072, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.121846] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1127.121846] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52af5b56-eb34-eb0d-eb90-e35d955cff28" [ 1127.121846] env[61728]: _type = "HttpNfcLease" [ 1127.121846] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1127.492895] env[61728]: DEBUG nova.compute.utils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1127.494394] env[61728]: DEBUG nova.compute.manager [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1127.494568] env[61728]: DEBUG nova.network.neutron [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1127.544731] env[61728]: DEBUG nova.policy [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3baf998a00e494ba33dab6a0a2a88e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '836843da5be34c649d9a48a83e658288', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1127.616381] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52beeba9-501b-3c3f-d1e7-3c7629dc5072, 'name': SearchDatastore_Task, 'duration_secs': 0.0142} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.620024] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.620324] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] dbe877e6-1754-4cf7-b454-abd88bfa713a/dbe877e6-1754-4cf7-b454-abd88bfa713a.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1127.621050] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c560427b-33e9-40f1-8d5e-80665aa624c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.626826] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1127.626826] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52af5b56-eb34-eb0d-eb90-e35d955cff28" [ 1127.626826] env[61728]: _type = "HttpNfcLease" [ 1127.626826] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1127.628736] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1127.628736] env[61728]: value = "task-464968" [ 1127.628736] env[61728]: _type = "Task" [ 1127.628736] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.638038] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464968, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.840300] env[61728]: DEBUG nova.network.neutron [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Successfully created port: e6562f3d-358f-482b-b0a1-b6a09c232c71 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1127.998476] env[61728]: DEBUG nova.compute.manager [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1128.126503] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1128.126503] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52af5b56-eb34-eb0d-eb90-e35d955cff28" [ 1128.126503] env[61728]: _type = "HttpNfcLease" [ 1128.126503] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1128.139542] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464968, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465755} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.139887] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] dbe877e6-1754-4cf7-b454-abd88bfa713a/dbe877e6-1754-4cf7-b454-abd88bfa713a.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1128.139977] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1128.142522] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f77c6a4f-e884-4204-865d-70aaeedcb4d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.149127] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1128.149127] env[61728]: value = "task-464969" [ 1128.149127] env[61728]: _type = "Task" [ 1128.149127] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.157242] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464969, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.226742] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-484d6357-b94c-4d15-af7f-4d07afe589f1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.233988] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e9df3f-6e7b-45d1-98d8-800e3f62c488 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.264036] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff841231-8b03-44f1-93c9-e0a8223bee4c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.271246] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7858e61b-cfac-4b94-a290-401156db7e20 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.289906] env[61728]: DEBUG nova.compute.provider_tree [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1128.629435] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1128.629435] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52af5b56-eb34-eb0d-eb90-e35d955cff28" [ 1128.629435] env[61728]: _type = "HttpNfcLease" [ 1128.629435] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1128.629435] env[61728]: DEBUG oslo_vmware.rw_handles [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1128.629435] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52af5b56-eb34-eb0d-eb90-e35d955cff28" [ 1128.629435] env[61728]: _type = "HttpNfcLease" [ 1128.629435] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1128.629435] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9ac820-446a-400c-9468-fd0ddec38b53 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.638525] env[61728]: DEBUG oslo_vmware.rw_handles [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5248af01-c248-7758-a021-3e8518e4857f/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1128.638620] env[61728]: DEBUG oslo_vmware.rw_handles [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5248af01-c248-7758-a021-3e8518e4857f/disk-0.vmdk. {{(pid=61728) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1128.717041] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0d611ec8-1f17-4329-b4af-8c1873f6146b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.722023] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464969, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068598} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.722023] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1128.722023] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-840cd5c1-6113-4320-b752-eb09796ba28f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.742529] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] dbe877e6-1754-4cf7-b454-abd88bfa713a/dbe877e6-1754-4cf7-b454-abd88bfa713a.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1128.742917] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34ba24b9-666e-4db2-af26-2cba521f1530 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.762572] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1128.762572] env[61728]: value = "task-464970" [ 1128.762572] env[61728]: _type = "Task" [ 1128.762572] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.770551] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464970, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.794784] env[61728]: DEBUG nova.scheduler.client.report [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1128.897671] env[61728]: DEBUG oslo_concurrency.lockutils [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.897913] env[61728]: DEBUG oslo_concurrency.lockutils [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.010077] env[61728]: DEBUG nova.compute.manager [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1129.276202] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464970, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.300072] env[61728]: DEBUG nova.compute.manager [req-671cd712-5840-460a-b433-b7a58d0a5dbd req-99e46671-f581-4d14-adcd-08abf78db55b service nova] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Received event network-vif-plugged-e6562f3d-358f-482b-b0a1-b6a09c232c71 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1129.300311] env[61728]: DEBUG oslo_concurrency.lockutils [req-671cd712-5840-460a-b433-b7a58d0a5dbd req-99e46671-f581-4d14-adcd-08abf78db55b service nova] Acquiring lock "2db96467-5fb4-49f4-954b-dbdd21e46232-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.300520] env[61728]: DEBUG oslo_concurrency.lockutils [req-671cd712-5840-460a-b433-b7a58d0a5dbd req-99e46671-f581-4d14-adcd-08abf78db55b service nova] Lock "2db96467-5fb4-49f4-954b-dbdd21e46232-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.300691] env[61728]: DEBUG oslo_concurrency.lockutils [req-671cd712-5840-460a-b433-b7a58d0a5dbd req-99e46671-f581-4d14-adcd-08abf78db55b service nova] Lock "2db96467-5fb4-49f4-954b-dbdd21e46232-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.300861] env[61728]: DEBUG nova.compute.manager [req-671cd712-5840-460a-b433-b7a58d0a5dbd req-99e46671-f581-4d14-adcd-08abf78db55b service nova] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] No waiting events found dispatching network-vif-plugged-e6562f3d-358f-482b-b0a1-b6a09c232c71 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1129.301212] env[61728]: WARNING nova.compute.manager [req-671cd712-5840-460a-b433-b7a58d0a5dbd req-99e46671-f581-4d14-adcd-08abf78db55b service nova] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Received unexpected event network-vif-plugged-e6562f3d-358f-482b-b0a1-b6a09c232c71 for instance with vm_state building and task_state spawning. [ 1129.302086] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.254s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.302651] env[61728]: DEBUG nova.compute.manager [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1129.403110] env[61728]: DEBUG nova.compute.utils [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1129.680590] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.681118] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.681321] env[61728]: DEBUG nova.compute.manager [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Going to confirm migration 8 {{(pid=61728) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1129.741172] env[61728]: DEBUG nova.virt.hardware [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1129.741534] env[61728]: DEBUG nova.virt.hardware [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1129.741725] env[61728]: DEBUG nova.virt.hardware [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1129.741935] env[61728]: DEBUG nova.virt.hardware [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1129.742269] env[61728]: DEBUG nova.virt.hardware [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1129.742507] env[61728]: DEBUG nova.virt.hardware [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1129.742760] env[61728]: DEBUG nova.virt.hardware [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1129.742981] env[61728]: DEBUG nova.virt.hardware [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1129.743316] env[61728]: DEBUG nova.virt.hardware [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1129.743545] env[61728]: DEBUG nova.virt.hardware [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1129.743737] env[61728]: DEBUG nova.virt.hardware [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1129.745257] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-792e09c4-baa6-4282-9d95-25d3777297d3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.758444] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8013a4bb-63b8-4bb8-b181-41c0c023fb9d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.789711] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464970, 'name': ReconfigVM_Task, 'duration_secs': 0.619319} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.791862] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Reconfigured VM instance instance-00000075 to attach disk [datastore1] dbe877e6-1754-4cf7-b454-abd88bfa713a/dbe877e6-1754-4cf7-b454-abd88bfa713a.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1129.792639] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-33ba58e0-438c-44c2-8495-a23623fd66dc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.799735] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1129.799735] env[61728]: value = "task-464971" [ 1129.799735] env[61728]: _type = "Task" [ 1129.799735] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.811063] env[61728]: DEBUG nova.compute.utils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1129.812644] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464971, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.814916] env[61728]: DEBUG nova.compute.manager [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1129.815111] env[61728]: DEBUG nova.network.neutron [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1129.857822] env[61728]: DEBUG nova.network.neutron [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Successfully updated port: e6562f3d-358f-482b-b0a1-b6a09c232c71 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1129.883982] env[61728]: DEBUG nova.policy [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '39df6cef3a6f430d92e15d68076e759f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '909c86730f554a04ab05dcad3a53e9dd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1129.908257] env[61728]: DEBUG oslo_concurrency.lockutils [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.168928] env[61728]: DEBUG nova.network.neutron [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Successfully created port: 7aee91de-db46-4071-8f9e-980bcfc29365 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1130.247245] env[61728]: DEBUG oslo_vmware.rw_handles [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Completed reading data from the image iterator. {{(pid=61728) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1130.247544] env[61728]: DEBUG oslo_vmware.rw_handles [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5248af01-c248-7758-a021-3e8518e4857f/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1130.248530] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1446c107-4227-44cf-b446-2c7cf41c661b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.256137] env[61728]: DEBUG oslo_vmware.rw_handles [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5248af01-c248-7758-a021-3e8518e4857f/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1130.256347] env[61728]: DEBUG oslo_vmware.rw_handles [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5248af01-c248-7758-a021-3e8518e4857f/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1130.256595] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-2859411d-7662-4c7e-af67-de87e7eea81c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.299775] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.300059] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.300268] env[61728]: DEBUG nova.network.neutron [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1130.300466] env[61728]: DEBUG nova.objects.instance [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lazy-loading 'info_cache' on Instance uuid 8c5b2380-bd14-451e-9612-2767c2a6adb2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.311036] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464971, 'name': Rename_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.318716] env[61728]: DEBUG nova.compute.manager [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1130.352306] env[61728]: DEBUG nova.compute.manager [req-7d32c66a-8161-41bd-af8d-956e659051e3 req-92fb4b40-8158-43de-afd2-b44e3d6a4843 service nova] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Received event network-changed-e6562f3d-358f-482b-b0a1-b6a09c232c71 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1130.352617] env[61728]: DEBUG nova.compute.manager [req-7d32c66a-8161-41bd-af8d-956e659051e3 req-92fb4b40-8158-43de-afd2-b44e3d6a4843 service nova] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Refreshing instance network info cache due to event network-changed-e6562f3d-358f-482b-b0a1-b6a09c232c71. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1130.352931] env[61728]: DEBUG oslo_concurrency.lockutils [req-7d32c66a-8161-41bd-af8d-956e659051e3 req-92fb4b40-8158-43de-afd2-b44e3d6a4843 service nova] Acquiring lock "refresh_cache-2db96467-5fb4-49f4-954b-dbdd21e46232" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.353228] env[61728]: DEBUG oslo_concurrency.lockutils [req-7d32c66a-8161-41bd-af8d-956e659051e3 req-92fb4b40-8158-43de-afd2-b44e3d6a4843 service nova] Acquired lock "refresh_cache-2db96467-5fb4-49f4-954b-dbdd21e46232" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.353474] env[61728]: DEBUG nova.network.neutron [req-7d32c66a-8161-41bd-af8d-956e659051e3 req-92fb4b40-8158-43de-afd2-b44e3d6a4843 service nova] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Refreshing network info cache for port e6562f3d-358f-482b-b0a1-b6a09c232c71 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1130.360365] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "refresh_cache-2db96467-5fb4-49f4-954b-dbdd21e46232" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.622479] env[61728]: DEBUG oslo_vmware.rw_handles [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5248af01-c248-7758-a021-3e8518e4857f/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1130.622775] env[61728]: INFO nova.virt.vmwareapi.images [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Downloaded image file data d9914484-11b1-4819-982a-67a108bd3ed1 [ 1130.623789] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82bfd848-c971-4ab2-8832-52436969ae5d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.638969] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c75b315d-b92b-4b87-a3ab-5686ce4ed020 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.748976] env[61728]: INFO nova.virt.vmwareapi.images [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] The imported VM was unregistered [ 1130.750991] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Caching image {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1130.751256] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Creating directory with path [datastore1] devstack-image-cache_base/d9914484-11b1-4819-982a-67a108bd3ed1 {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1130.751543] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-81077c4f-05c4-4221-89b5-133df85aa504 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.762393] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Created directory with path [datastore1] devstack-image-cache_base/d9914484-11b1-4819-982a-67a108bd3ed1 {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1130.762589] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_e1bbc824-f565-4bca-8a9a-f6cb7551993c/OSTACK_IMG_e1bbc824-f565-4bca-8a9a-f6cb7551993c.vmdk to [datastore1] devstack-image-cache_base/d9914484-11b1-4819-982a-67a108bd3ed1/d9914484-11b1-4819-982a-67a108bd3ed1.vmdk. {{(pid=61728) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1130.762859] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-a5774b50-98a7-4af5-a75a-f2b749bf8ea6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.768971] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1130.768971] env[61728]: value = "task-464973" [ 1130.768971] env[61728]: _type = "Task" [ 1130.768971] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.777014] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464973, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.813671] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464971, 'name': Rename_Task, 'duration_secs': 0.799174} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.813986] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1130.814311] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f384a1f-c2b5-4ed3-b157-d51c24b7908d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.825527] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1130.825527] env[61728]: value = "task-464974" [ 1130.825527] env[61728]: _type = "Task" [ 1130.825527] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.834025] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464974, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.886151] env[61728]: DEBUG nova.network.neutron [req-7d32c66a-8161-41bd-af8d-956e659051e3 req-92fb4b40-8158-43de-afd2-b44e3d6a4843 service nova] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1130.966525] env[61728]: DEBUG nova.network.neutron [req-7d32c66a-8161-41bd-af8d-956e659051e3 req-92fb4b40-8158-43de-afd2-b44e3d6a4843 service nova] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.975175] env[61728]: DEBUG oslo_concurrency.lockutils [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.975498] env[61728]: DEBUG oslo_concurrency.lockutils [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.975792] env[61728]: INFO nova.compute.manager [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Attaching volume 4b08464d-0026-4ba9-a28e-88e941c8f364 to /dev/sdb [ 1131.017106] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36647633-bfb8-4a43-bb94-b3365ee4d8dc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.026503] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0637bbc-4210-456d-a347-9c660d19ca6f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.041240] env[61728]: DEBUG nova.virt.block_device [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Updating existing volume attachment record: cfb41685-0a06-4bc0-a648-9239af539605 {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1131.278460] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464973, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.327764] env[61728]: DEBUG nova.compute.manager [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1131.340758] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464974, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.359206] env[61728]: DEBUG nova.virt.hardware [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1131.359675] env[61728]: DEBUG nova.virt.hardware [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1131.359974] env[61728]: DEBUG nova.virt.hardware [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1131.360339] env[61728]: DEBUG nova.virt.hardware [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1131.360603] env[61728]: DEBUG nova.virt.hardware [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1131.360890] env[61728]: DEBUG nova.virt.hardware [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1131.361309] env[61728]: DEBUG nova.virt.hardware [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1131.361614] env[61728]: DEBUG nova.virt.hardware [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1131.361973] env[61728]: DEBUG nova.virt.hardware [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1131.362300] env[61728]: DEBUG nova.virt.hardware [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1131.362653] env[61728]: DEBUG nova.virt.hardware [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1131.364147] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c35cef-7fa1-4d6a-a320-dc3e5a9eacfd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.375133] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eee21cb-a84f-4b6b-a6f3-93486f818692 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.471537] env[61728]: DEBUG oslo_concurrency.lockutils [req-7d32c66a-8161-41bd-af8d-956e659051e3 req-92fb4b40-8158-43de-afd2-b44e3d6a4843 service nova] Releasing lock "refresh_cache-2db96467-5fb4-49f4-954b-dbdd21e46232" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.471537] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "refresh_cache-2db96467-5fb4-49f4-954b-dbdd21e46232" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.471537] env[61728]: DEBUG nova.network.neutron [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1131.690238] env[61728]: DEBUG nova.network.neutron [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance_info_cache with network_info: [{"id": "4660e4da-30e8-4e51-9aea-a9375732537f", "address": "fa:16:3e:e1:e9:0c", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4660e4da-30", "ovs_interfaceid": "4660e4da-30e8-4e51-9aea-a9375732537f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.781103] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464973, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.845026] env[61728]: DEBUG oslo_vmware.api [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464974, 'name': PowerOnVM_Task, 'duration_secs': 0.667025} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.845026] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1131.845026] env[61728]: DEBUG nova.compute.manager [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1131.845026] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee62e6f7-9f88-47d0-8618-cb5df69921b5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.885784] env[61728]: DEBUG nova.network.neutron [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Successfully updated port: 7aee91de-db46-4071-8f9e-980bcfc29365 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1132.006810] env[61728]: DEBUG nova.network.neutron [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1132.162760] env[61728]: DEBUG nova.network.neutron [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Updating instance_info_cache with network_info: [{"id": "e6562f3d-358f-482b-b0a1-b6a09c232c71", "address": "fa:16:3e:8a:36:db", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6562f3d-35", "ovs_interfaceid": "e6562f3d-358f-482b-b0a1-b6a09c232c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.193916] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-8c5b2380-bd14-451e-9612-2767c2a6adb2" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.194248] env[61728]: DEBUG nova.objects.instance [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lazy-loading 'migration_context' on Instance uuid 8c5b2380-bd14-451e-9612-2767c2a6adb2 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1132.283309] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464973, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.366977] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.367424] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.367713] env[61728]: DEBUG nova.objects.instance [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61728) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1132.385627] env[61728]: DEBUG nova.compute.manager [req-6250e709-a571-4fba-a7cb-6526d1787351 req-ad039ded-58a8-4a96-9622-b7447f4aca33 service nova] [instance: c92e595d-c686-4810-bb39-d98de4141121] Received event network-vif-plugged-7aee91de-db46-4071-8f9e-980bcfc29365 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1132.385942] env[61728]: DEBUG oslo_concurrency.lockutils [req-6250e709-a571-4fba-a7cb-6526d1787351 req-ad039ded-58a8-4a96-9622-b7447f4aca33 service nova] Acquiring lock "c92e595d-c686-4810-bb39-d98de4141121-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.386205] env[61728]: DEBUG oslo_concurrency.lockutils [req-6250e709-a571-4fba-a7cb-6526d1787351 req-ad039ded-58a8-4a96-9622-b7447f4aca33 service nova] Lock "c92e595d-c686-4810-bb39-d98de4141121-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.386441] env[61728]: DEBUG oslo_concurrency.lockutils [req-6250e709-a571-4fba-a7cb-6526d1787351 req-ad039ded-58a8-4a96-9622-b7447f4aca33 service nova] Lock "c92e595d-c686-4810-bb39-d98de4141121-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.386642] env[61728]: DEBUG nova.compute.manager [req-6250e709-a571-4fba-a7cb-6526d1787351 req-ad039ded-58a8-4a96-9622-b7447f4aca33 service nova] [instance: c92e595d-c686-4810-bb39-d98de4141121] No waiting events found dispatching network-vif-plugged-7aee91de-db46-4071-8f9e-980bcfc29365 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1132.386781] env[61728]: WARNING nova.compute.manager [req-6250e709-a571-4fba-a7cb-6526d1787351 req-ad039ded-58a8-4a96-9622-b7447f4aca33 service nova] [instance: c92e595d-c686-4810-bb39-d98de4141121] Received unexpected event network-vif-plugged-7aee91de-db46-4071-8f9e-980bcfc29365 for instance with vm_state building and task_state spawning. [ 1132.386986] env[61728]: DEBUG nova.compute.manager [req-6250e709-a571-4fba-a7cb-6526d1787351 req-ad039ded-58a8-4a96-9622-b7447f4aca33 service nova] [instance: c92e595d-c686-4810-bb39-d98de4141121] Received event network-changed-7aee91de-db46-4071-8f9e-980bcfc29365 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1132.387209] env[61728]: DEBUG nova.compute.manager [req-6250e709-a571-4fba-a7cb-6526d1787351 req-ad039ded-58a8-4a96-9622-b7447f4aca33 service nova] [instance: c92e595d-c686-4810-bb39-d98de4141121] Refreshing instance network info cache due to event network-changed-7aee91de-db46-4071-8f9e-980bcfc29365. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1132.387404] env[61728]: DEBUG oslo_concurrency.lockutils [req-6250e709-a571-4fba-a7cb-6526d1787351 req-ad039ded-58a8-4a96-9622-b7447f4aca33 service nova] Acquiring lock "refresh_cache-c92e595d-c686-4810-bb39-d98de4141121" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.387552] env[61728]: DEBUG oslo_concurrency.lockutils [req-6250e709-a571-4fba-a7cb-6526d1787351 req-ad039ded-58a8-4a96-9622-b7447f4aca33 service nova] Acquired lock "refresh_cache-c92e595d-c686-4810-bb39-d98de4141121" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.387716] env[61728]: DEBUG nova.network.neutron [req-6250e709-a571-4fba-a7cb-6526d1787351 req-ad039ded-58a8-4a96-9622-b7447f4aca33 service nova] [instance: c92e595d-c686-4810-bb39-d98de4141121] Refreshing network info cache for port 7aee91de-db46-4071-8f9e-980bcfc29365 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1132.389531] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "refresh_cache-c92e595d-c686-4810-bb39-d98de4141121" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.512318] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e413d2-fd60-5852-0a38-b7449bd5f849/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1132.513387] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a1174b-9e79-4bc7-b610-4b16521d392d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.520677] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e413d2-fd60-5852-0a38-b7449bd5f849/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1132.520872] env[61728]: ERROR oslo_vmware.rw_handles [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e413d2-fd60-5852-0a38-b7449bd5f849/disk-0.vmdk due to incomplete transfer. [ 1132.521157] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1058a8a8-46f9-49bd-a31e-849e1454cd36 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.529181] env[61728]: DEBUG oslo_vmware.rw_handles [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e413d2-fd60-5852-0a38-b7449bd5f849/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1132.529411] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Uploaded image df2f6391-a88d-4bb9-ada0-dd01a6acec62 to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1132.531866] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1132.532254] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-58e9204c-95df-4e2d-afc2-aeca7400fd3e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.539869] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1132.539869] env[61728]: value = "task-464976" [ 1132.539869] env[61728]: _type = "Task" [ 1132.539869] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.552144] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464976, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.666211] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "refresh_cache-2db96467-5fb4-49f4-954b-dbdd21e46232" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.666211] env[61728]: DEBUG nova.compute.manager [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Instance network_info: |[{"id": "e6562f3d-358f-482b-b0a1-b6a09c232c71", "address": "fa:16:3e:8a:36:db", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6562f3d-35", "ovs_interfaceid": "e6562f3d-358f-482b-b0a1-b6a09c232c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1132.666991] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:36:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9732690c-bdcf-4e6f-9a32-42c196333eb8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e6562f3d-358f-482b-b0a1-b6a09c232c71', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1132.675247] env[61728]: DEBUG oslo.service.loopingcall [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1132.676704] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1132.679226] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d2e6f86-9f4d-4fd4-9233-372ffa2f4a7c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.697206] env[61728]: DEBUG nova.objects.base [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Object Instance<8c5b2380-bd14-451e-9612-2767c2a6adb2> lazy-loaded attributes: info_cache,migration_context {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1132.698703] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-135aaf56-4ef8-4e94-aa29-fc596f05946b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.703119] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1132.703119] env[61728]: value = "task-464977" [ 1132.703119] env[61728]: _type = "Task" [ 1132.703119] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.721686] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea40118c-bb99-40b5-aa46-724ac56ead5e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.727517] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464977, 'name': CreateVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.731515] env[61728]: DEBUG oslo_vmware.api [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1132.731515] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c4b63c-ce1c-4042-3351-10c8362beca4" [ 1132.731515] env[61728]: _type = "Task" [ 1132.731515] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.741130] env[61728]: DEBUG oslo_vmware.api [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c4b63c-ce1c-4042-3351-10c8362beca4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.782009] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464973, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.056060] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464976, 'name': Destroy_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.120291] env[61728]: DEBUG nova.network.neutron [req-6250e709-a571-4fba-a7cb-6526d1787351 req-ad039ded-58a8-4a96-9622-b7447f4aca33 service nova] [instance: c92e595d-c686-4810-bb39-d98de4141121] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1133.195473] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquiring lock "dbe877e6-1754-4cf7-b454-abd88bfa713a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.195756] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Lock "dbe877e6-1754-4cf7-b454-abd88bfa713a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.196028] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquiring lock "dbe877e6-1754-4cf7-b454-abd88bfa713a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.196357] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Lock "dbe877e6-1754-4cf7-b454-abd88bfa713a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.196571] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Lock "dbe877e6-1754-4cf7-b454-abd88bfa713a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.199393] env[61728]: INFO nova.compute.manager [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Terminating instance [ 1133.201307] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquiring lock "refresh_cache-dbe877e6-1754-4cf7-b454-abd88bfa713a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.201516] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquired lock "refresh_cache-dbe877e6-1754-4cf7-b454-abd88bfa713a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.201781] env[61728]: DEBUG nova.network.neutron [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1133.210144] env[61728]: DEBUG nova.network.neutron [req-6250e709-a571-4fba-a7cb-6526d1787351 req-ad039ded-58a8-4a96-9622-b7447f4aca33 service nova] [instance: c92e595d-c686-4810-bb39-d98de4141121] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.214970] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464977, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.242375] env[61728]: DEBUG oslo_vmware.api [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c4b63c-ce1c-4042-3351-10c8362beca4, 'name': SearchDatastore_Task, 'duration_secs': 0.045989} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.242757] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.280911] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464973, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.377317] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8ddb5bb1-a22e-41cc-9d73-e3b4c95681e6 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.378530] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.136s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.550295] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464976, 'name': Destroy_Task, 'duration_secs': 0.867326} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.550584] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Destroyed the VM [ 1133.550826] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1133.551108] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b6e1ca95-b75a-4e66-9725-97513901d948 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.558591] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1133.558591] env[61728]: value = "task-464979" [ 1133.558591] env[61728]: _type = "Task" [ 1133.558591] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.567020] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464979, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.715329] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464977, 'name': CreateVM_Task, 'duration_secs': 0.756276} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.715958] env[61728]: DEBUG oslo_concurrency.lockutils [req-6250e709-a571-4fba-a7cb-6526d1787351 req-ad039ded-58a8-4a96-9622-b7447f4aca33 service nova] Releasing lock "refresh_cache-c92e595d-c686-4810-bb39-d98de4141121" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.716141] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1133.716635] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquired lock "refresh_cache-c92e595d-c686-4810-bb39-d98de4141121" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.716635] env[61728]: DEBUG nova.network.neutron [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1133.717961] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.718577] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.719145] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1133.719366] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a3fbed6-3505-4dd0-a7c1-892cc55340cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.724814] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1133.724814] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e029cb-8565-8cf7-41c9-3e27a64b616c" [ 1133.724814] env[61728]: _type = "Task" [ 1133.724814] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.725679] env[61728]: DEBUG nova.network.neutron [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1133.737554] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52e029cb-8565-8cf7-41c9-3e27a64b616c, 'name': SearchDatastore_Task, 'duration_secs': 0.008989} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.737804] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.738535] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1133.738535] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.738535] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.738728] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1133.739691] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d62ef9d-45ee-403e-9bbd-eb1baf59b0e3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.746104] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1133.746295] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1133.747044] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f75a917-92c1-4fb0-b591-2f6b8bfeb719 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.752027] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1133.752027] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f5c8c8-f80e-3902-92d3-1dfff9fabd4b" [ 1133.752027] env[61728]: _type = "Task" [ 1133.752027] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.759738] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f5c8c8-f80e-3902-92d3-1dfff9fabd4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.782235] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464973, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.51487} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.783108] env[61728]: DEBUG nova.network.neutron [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.784852] env[61728]: INFO nova.virt.vmwareapi.ds_util [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_e1bbc824-f565-4bca-8a9a-f6cb7551993c/OSTACK_IMG_e1bbc824-f565-4bca-8a9a-f6cb7551993c.vmdk to [datastore1] devstack-image-cache_base/d9914484-11b1-4819-982a-67a108bd3ed1/d9914484-11b1-4819-982a-67a108bd3ed1.vmdk. [ 1133.785136] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Cleaning up location [datastore1] OSTACK_IMG_e1bbc824-f565-4bca-8a9a-f6cb7551993c {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1133.785336] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_e1bbc824-f565-4bca-8a9a-f6cb7551993c {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1133.785592] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-83f62b6a-f7d1-4d81-a829-870bf4fa6111 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.792581] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1133.792581] env[61728]: value = "task-464980" [ 1133.792581] env[61728]: _type = "Task" [ 1133.792581] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.802364] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464980, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.036014] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f18cd7-04e4-4e30-9267-2d00d925bf96 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.044069] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78cbabfe-ac6e-4fed-bccc-e4c2e760346f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.078503] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18533911-7f68-4186-b442-b74235c3d8c9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.086482] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464979, 'name': RemoveSnapshot_Task, 'duration_secs': 0.418023} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.088731] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1134.089048] env[61728]: DEBUG nova.compute.manager [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1134.089882] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24d4bc1-ea89-4fff-8f88-74d9607cdf67 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.093288] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d981acef-0e6a-498f-961d-ae8c70b65f4d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.106886] env[61728]: DEBUG nova.compute.provider_tree [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1134.249879] env[61728]: DEBUG nova.network.neutron [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1134.262613] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f5c8c8-f80e-3902-92d3-1dfff9fabd4b, 'name': SearchDatastore_Task, 'duration_secs': 0.008259} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.265433] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d360f38d-6896-42ec-b17e-96e151d0382c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.270803] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1134.270803] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5249f175-2e3f-c958-3800-3372c7d144f0" [ 1134.270803] env[61728]: _type = "Task" [ 1134.270803] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.278803] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5249f175-2e3f-c958-3800-3372c7d144f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.285425] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Releasing lock "refresh_cache-dbe877e6-1754-4cf7-b454-abd88bfa713a" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.285919] env[61728]: DEBUG nova.compute.manager [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1134.286173] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1134.287007] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54545d51-19ff-42ee-9e69-d89b949b6e34 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.298683] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1134.299265] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3674ed6a-8326-494b-8e99-12c1897e5616 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.304173] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464980, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034175} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.304447] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1134.304639] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d9914484-11b1-4819-982a-67a108bd3ed1/d9914484-11b1-4819-982a-67a108bd3ed1.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.304891] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d9914484-11b1-4819-982a-67a108bd3ed1/d9914484-11b1-4819-982a-67a108bd3ed1.vmdk to [datastore1] 77b8f509-17e7-4c25-a43d-1655c16ce8d1/77b8f509-17e7-4c25-a43d-1655c16ce8d1.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1134.305964] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7857b9ec-3055-4345-be46-e073945e5c23 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.307932] env[61728]: DEBUG oslo_vmware.api [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1134.307932] env[61728]: value = "task-464981" [ 1134.307932] env[61728]: _type = "Task" [ 1134.307932] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.315208] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1134.315208] env[61728]: value = "task-464982" [ 1134.315208] env[61728]: _type = "Task" [ 1134.315208] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.321048] env[61728]: DEBUG oslo_vmware.api [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464981, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.325797] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464982, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.387274] env[61728]: DEBUG nova.network.neutron [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Updating instance_info_cache with network_info: [{"id": "7aee91de-db46-4071-8f9e-980bcfc29365", "address": "fa:16:3e:65:1f:da", "network": {"id": "4bb51519-042f-44ea-9325-a479be9790d3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1499510217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "909c86730f554a04ab05dcad3a53e9dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7aee91de-db", "ovs_interfaceid": "7aee91de-db46-4071-8f9e-980bcfc29365", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.617892] env[61728]: INFO nova.compute.manager [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Shelve offloading [ 1134.620212] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1134.620514] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f9f4499-65ce-4c93-9cd1-32af9b9cf66a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.629325] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1134.629325] env[61728]: value = "task-464983" [ 1134.629325] env[61728]: _type = "Task" [ 1134.629325] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.635037] env[61728]: ERROR nova.scheduler.client.report [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [req-6356ec7e-e412-432a-9281-7c789dead084] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6356ec7e-e412-432a-9281-7c789dead084"}]} [ 1134.643560] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1134.643560] env[61728]: DEBUG nova.compute.manager [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1134.644160] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567e619c-4a4e-4dbe-9d40-4e7484f1c2ba {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.650479] env[61728]: DEBUG nova.scheduler.client.report [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1134.654084] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.654084] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.654331] env[61728]: DEBUG nova.network.neutron [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1134.668056] env[61728]: DEBUG nova.scheduler.client.report [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1134.668374] env[61728]: DEBUG nova.compute.provider_tree [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1134.681732] env[61728]: DEBUG nova.scheduler.client.report [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1134.701478] env[61728]: DEBUG nova.scheduler.client.report [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1134.783291] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5249f175-2e3f-c958-3800-3372c7d144f0, 'name': SearchDatastore_Task, 'duration_secs': 0.008971} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.786031] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.786332] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2db96467-5fb4-49f4-954b-dbdd21e46232/2db96467-5fb4-49f4-954b-dbdd21e46232.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1134.786816] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1dffe42-7652-480e-b18f-b9aefb305e3c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.793888] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1134.793888] env[61728]: value = "task-464984" [ 1134.793888] env[61728]: _type = "Task" [ 1134.793888] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.805348] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464984, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.823225] env[61728]: DEBUG oslo_vmware.api [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464981, 'name': PowerOffVM_Task, 'duration_secs': 0.123176} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.826521] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1134.827410] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1134.827410] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a15d6bc4-f95b-4763-8623-643174086793 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.831843] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464982, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.854304] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1134.854536] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1134.854742] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Deleting the datastore file [datastore1] dbe877e6-1754-4cf7-b454-abd88bfa713a {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1134.855190] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7e4d5c0-cc49-4f13-b3cd-40d662c59e70 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.862995] env[61728]: DEBUG oslo_vmware.api [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for the task: (returnval){ [ 1134.862995] env[61728]: value = "task-464986" [ 1134.862995] env[61728]: _type = "Task" [ 1134.862995] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.864789] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4244682-fd7d-4250-8577-1f0ecfa662cc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.875507] env[61728]: DEBUG oslo_vmware.api [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464986, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.877942] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5f83e9-0d51-419a-ba68-1075a302d25c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.909266] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Releasing lock "refresh_cache-c92e595d-c686-4810-bb39-d98de4141121" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.909591] env[61728]: DEBUG nova.compute.manager [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Instance network_info: |[{"id": "7aee91de-db46-4071-8f9e-980bcfc29365", "address": "fa:16:3e:65:1f:da", "network": {"id": "4bb51519-042f-44ea-9325-a479be9790d3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1499510217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "909c86730f554a04ab05dcad3a53e9dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7aee91de-db", "ovs_interfaceid": "7aee91de-db46-4071-8f9e-980bcfc29365", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1134.910369] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:1f:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '257e5ea7-8b80-4301-9900-a754f1fe2031', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7aee91de-db46-4071-8f9e-980bcfc29365', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1134.918063] env[61728]: DEBUG oslo.service.loopingcall [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1134.918829] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b1720b-dd4e-48e8-9982-d213c37afa9b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.921613] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c92e595d-c686-4810-bb39-d98de4141121] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1134.921896] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00f9de69-55ca-431c-9da9-faf65712ea13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.941675] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e54d611-f575-4a57-b470-80b3268ad92a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.946593] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1134.946593] env[61728]: value = "task-464987" [ 1134.946593] env[61728]: _type = "Task" [ 1134.946593] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.957979] env[61728]: DEBUG nova.compute.provider_tree [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1134.964286] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464987, 'name': CreateVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.095225] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.095623] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.305714] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464984, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.326753] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464982, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.378527] env[61728]: DEBUG oslo_vmware.api [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464986, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.409285] env[61728]: DEBUG nova.network.neutron [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updating instance_info_cache with network_info: [{"id": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "address": "fa:16:3e:f1:65:32", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04bc3a5-d3", "ovs_interfaceid": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.457316] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464987, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.492161] env[61728]: DEBUG nova.scheduler.client.report [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 183 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1135.492490] env[61728]: DEBUG nova.compute.provider_tree [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 183 to 184 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1135.492684] env[61728]: DEBUG nova.compute.provider_tree [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1135.591958] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1135.592259] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122229', 'volume_id': '4b08464d-0026-4ba9-a28e-88e941c8f364', 'name': 'volume-4b08464d-0026-4ba9-a28e-88e941c8f364', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c1ed0e8b-5cf9-4ff0-931e-fadc3432a125', 'attached_at': '', 'detached_at': '', 'volume_id': '4b08464d-0026-4ba9-a28e-88e941c8f364', 'serial': '4b08464d-0026-4ba9-a28e-88e941c8f364'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1135.593264] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f05433-e67c-4fe6-a057-3e1863e5bbcc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.599488] env[61728]: INFO nova.compute.manager [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Detaching volume 0136be7a-12a6-43a4-847f-203c70cf8dc8 [ 1135.614196] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-877c4ee0-9a9e-448c-a0a5-178f33af2302 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.643049] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] volume-4b08464d-0026-4ba9-a28e-88e941c8f364/volume-4b08464d-0026-4ba9-a28e-88e941c8f364.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1135.643886] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-389dee7c-4faf-4e68-ac0f-1495edf3601c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.657793] env[61728]: INFO nova.virt.block_device [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Attempting to driver detach volume 0136be7a-12a6-43a4-847f-203c70cf8dc8 from mountpoint /dev/sdb [ 1135.658066] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1135.658259] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122212', 'volume_id': '0136be7a-12a6-43a4-847f-203c70cf8dc8', 'name': 'volume-0136be7a-12a6-43a4-847f-203c70cf8dc8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd55beb02-00fc-4df6-a239-9e5d776a7c18', 'attached_at': '', 'detached_at': '', 'volume_id': '0136be7a-12a6-43a4-847f-203c70cf8dc8', 'serial': '0136be7a-12a6-43a4-847f-203c70cf8dc8'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1135.659159] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd6ee28-2563-47ee-9fa7-2b75e4b4cc71 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.682467] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b130569b-b41c-4e0b-8eb4-f2d22f38c3e5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.685338] env[61728]: DEBUG oslo_vmware.api [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1135.685338] env[61728]: value = "task-464988" [ 1135.685338] env[61728]: _type = "Task" [ 1135.685338] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.691813] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b8b16f-5741-419a-96e5-7042e23655a7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.700961] env[61728]: DEBUG oslo_vmware.api [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464988, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.717624] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b62f6ef-2566-4b0e-9b81-0420a9ba4c33 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.734825] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] The volume has not been displaced from its original location: [datastore1] volume-0136be7a-12a6-43a4-847f-203c70cf8dc8/volume-0136be7a-12a6-43a4-847f-203c70cf8dc8.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1135.740465] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Reconfiguring VM instance instance-00000069 to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1135.740897] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-177e7539-7dca-4859-b083-5e20e3af4d3c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.762416] env[61728]: DEBUG oslo_vmware.api [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1135.762416] env[61728]: value = "task-464989" [ 1135.762416] env[61728]: _type = "Task" [ 1135.762416] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.771041] env[61728]: DEBUG oslo_vmware.api [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464989, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.804967] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464984, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.828790] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464982, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.876481] env[61728]: DEBUG oslo_vmware.api [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464986, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.912689] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.957982] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464987, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.186677] env[61728]: DEBUG nova.compute.manager [req-e5f047c5-efb6-4635-819a-ba11e779a381 req-5ce805d7-2f7b-47e2-815a-416540edc8ce service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Received event network-vif-unplugged-d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1136.186948] env[61728]: DEBUG oslo_concurrency.lockutils [req-e5f047c5-efb6-4635-819a-ba11e779a381 req-5ce805d7-2f7b-47e2-815a-416540edc8ce service nova] Acquiring lock "532507fa-996d-45c7-bf70-f09c93be79ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.187306] env[61728]: DEBUG oslo_concurrency.lockutils [req-e5f047c5-efb6-4635-819a-ba11e779a381 req-5ce805d7-2f7b-47e2-815a-416540edc8ce service nova] Lock "532507fa-996d-45c7-bf70-f09c93be79ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.187680] env[61728]: DEBUG oslo_concurrency.lockutils [req-e5f047c5-efb6-4635-819a-ba11e779a381 req-5ce805d7-2f7b-47e2-815a-416540edc8ce service nova] Lock "532507fa-996d-45c7-bf70-f09c93be79ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.187760] env[61728]: DEBUG nova.compute.manager [req-e5f047c5-efb6-4635-819a-ba11e779a381 req-5ce805d7-2f7b-47e2-815a-416540edc8ce service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] No waiting events found dispatching network-vif-unplugged-d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1136.187898] env[61728]: WARNING nova.compute.manager [req-e5f047c5-efb6-4635-819a-ba11e779a381 req-5ce805d7-2f7b-47e2-815a-416540edc8ce service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Received unexpected event network-vif-unplugged-d04bc3a5-d378-4e1e-a145-ec523bcc442a for instance with vm_state shelved and task_state shelving_offloading. [ 1136.199011] env[61728]: DEBUG oslo_vmware.api [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464988, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.274067] env[61728]: DEBUG oslo_vmware.api [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464989, 'name': ReconfigVM_Task, 'duration_secs': 0.264101} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.274739] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Reconfigured VM instance instance-00000069 to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1136.279989] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8adccb38-d156-46b3-b317-1b820b7e4b46 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.291297] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1136.292210] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414980ce-2bc4-49f3-8787-9819bb20f939 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.300899] env[61728]: DEBUG oslo_vmware.api [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1136.300899] env[61728]: value = "task-464990" [ 1136.300899] env[61728]: _type = "Task" [ 1136.300899] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.304420] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464984, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.309250] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1136.309567] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-204b3bcd-536c-496e-9b53-52b202d4d978 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.316458] env[61728]: DEBUG oslo_vmware.api [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464990, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.326157] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464982, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.373163] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1136.373365] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1136.373529] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleting the datastore file [datastore1] 532507fa-996d-45c7-bf70-f09c93be79ed {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1136.376746] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1ac86c6e-1443-47a3-9750-69fd3bb73244 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.378678] env[61728]: DEBUG oslo_vmware.api [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Task: {'id': task-464986, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.074122} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.378927] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1136.379138] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1136.379327] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1136.379510] env[61728]: INFO nova.compute.manager [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Took 2.09 seconds to destroy the instance on the hypervisor. [ 1136.379758] env[61728]: DEBUG oslo.service.loopingcall [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1136.380371] env[61728]: DEBUG nova.compute.manager [-] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1136.380473] env[61728]: DEBUG nova.network.neutron [-] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1136.383954] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1136.383954] env[61728]: value = "task-464992" [ 1136.383954] env[61728]: _type = "Task" [ 1136.383954] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.391882] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464992, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.396512] env[61728]: DEBUG nova.network.neutron [-] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1136.458754] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464987, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.504071] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 3.125s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.699207] env[61728]: DEBUG oslo_vmware.api [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464988, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.805886] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464984, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.817412] env[61728]: DEBUG oslo_vmware.api [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464990, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.827457] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464982, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.457311} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.827793] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d9914484-11b1-4819-982a-67a108bd3ed1/d9914484-11b1-4819-982a-67a108bd3ed1.vmdk to [datastore1] 77b8f509-17e7-4c25-a43d-1655c16ce8d1/77b8f509-17e7-4c25-a43d-1655c16ce8d1.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1136.828667] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30730175-5255-4182-bb24-df6eb119934c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.853212] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 77b8f509-17e7-4c25-a43d-1655c16ce8d1/77b8f509-17e7-4c25-a43d-1655c16ce8d1.vmdk or device None with type streamOptimized {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1136.853687] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5228b323-bcda-4690-a26b-62dab710efcd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.876612] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1136.876612] env[61728]: value = "task-464993" [ 1136.876612] env[61728]: _type = "Task" [ 1136.876612] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.886498] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464993, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.894806] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464992, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.899304] env[61728]: DEBUG nova.network.neutron [-] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.960732] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-464987, 'name': CreateVM_Task, 'duration_secs': 1.869781} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.960939] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c92e595d-c686-4810-bb39-d98de4141121] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1136.961700] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1136.961922] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.962343] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1136.962638] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-336d56d0-991a-41dc-aa4a-db238ba9e28a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.967759] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1136.967759] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524f281b-2848-6099-8f81-ddaecb741740" [ 1136.967759] env[61728]: _type = "Task" [ 1136.967759] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.976501] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524f281b-2848-6099-8f81-ddaecb741740, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.063990] env[61728]: INFO nova.scheduler.client.report [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleted allocation for migration 915164ff-8e6f-4c15-862f-656699649a53 [ 1137.199277] env[61728]: DEBUG oslo_vmware.api [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464988, 'name': ReconfigVM_Task, 'duration_secs': 1.443153} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.199620] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Reconfigured VM instance instance-00000072 to attach disk [datastore1] volume-4b08464d-0026-4ba9-a28e-88e941c8f364/volume-4b08464d-0026-4ba9-a28e-88e941c8f364.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1137.204778] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f521eaa-0cbb-4903-9f14-ced213afd240 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.220850] env[61728]: DEBUG oslo_vmware.api [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1137.220850] env[61728]: value = "task-464994" [ 1137.220850] env[61728]: _type = "Task" [ 1137.220850] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.228556] env[61728]: DEBUG oslo_vmware.api [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464994, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.305966] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464984, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.164599} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.309124] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 2db96467-5fb4-49f4-954b-dbdd21e46232/2db96467-5fb4-49f4-954b-dbdd21e46232.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1137.309348] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1137.309599] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fab3e374-1195-49bc-b9fc-7b99eed16963 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.316016] env[61728]: DEBUG oslo_vmware.api [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464990, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.317243] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1137.317243] env[61728]: value = "task-464995" [ 1137.317243] env[61728]: _type = "Task" [ 1137.317243] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.324523] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464995, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.386074] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464993, 'name': ReconfigVM_Task, 'duration_secs': 0.385893} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.386428] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 77b8f509-17e7-4c25-a43d-1655c16ce8d1/77b8f509-17e7-4c25-a43d-1655c16ce8d1.vmdk or device None with type streamOptimized {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1137.389575] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-46278b1d-3214-4c7c-8b75-3bcb8b582ae8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.396685] env[61728]: DEBUG oslo_vmware.api [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-464992, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.665078} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.397776] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1137.397962] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1137.398165] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1137.400054] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1137.400054] env[61728]: value = "task-464996" [ 1137.400054] env[61728]: _type = "Task" [ 1137.400054] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.403068] env[61728]: INFO nova.compute.manager [-] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Took 1.02 seconds to deallocate network for instance. [ 1137.410663] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464996, 'name': Rename_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.418862] env[61728]: INFO nova.scheduler.client.report [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleted allocations for instance 532507fa-996d-45c7-bf70-f09c93be79ed [ 1137.477920] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]524f281b-2848-6099-8f81-ddaecb741740, 'name': SearchDatastore_Task, 'duration_secs': 0.009211} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.478285] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.478501] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1137.478740] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.478897] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.479100] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1137.479396] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13c7b06b-882d-48ca-89f9-fba5761949a5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.486410] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1137.486590] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1137.487286] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0039a7dc-4ad8-48c6-9445-37f48f8e730d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.492104] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1137.492104] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528cc1a8-6651-18c0-a1d7-fade2928b28f" [ 1137.492104] env[61728]: _type = "Task" [ 1137.492104] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.499284] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528cc1a8-6651-18c0-a1d7-fade2928b28f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.570342] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cf23060d-0d39-4bf4-aa54-eb2f7f1b3d43 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.889s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.730391] env[61728]: DEBUG oslo_vmware.api [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-464994, 'name': ReconfigVM_Task, 'duration_secs': 0.14634} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.730719] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122229', 'volume_id': '4b08464d-0026-4ba9-a28e-88e941c8f364', 'name': 'volume-4b08464d-0026-4ba9-a28e-88e941c8f364', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c1ed0e8b-5cf9-4ff0-931e-fadc3432a125', 'attached_at': '', 'detached_at': '', 'volume_id': '4b08464d-0026-4ba9-a28e-88e941c8f364', 'serial': '4b08464d-0026-4ba9-a28e-88e941c8f364'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1137.816496] env[61728]: DEBUG oslo_vmware.api [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464990, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.825376] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464995, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.120781} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.825665] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1137.826483] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6df78b-0ceb-406f-a1a5-4535c8adb5b8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.847763] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Reconfiguring VM instance instance-00000076 to attach disk [datastore1] 2db96467-5fb4-49f4-954b-dbdd21e46232/2db96467-5fb4-49f4-954b-dbdd21e46232.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1137.849119] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-401eb6ac-6b12-4f4b-b852-0c005c912c68 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.871371] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1137.871371] env[61728]: value = "task-464997" [ 1137.871371] env[61728]: _type = "Task" [ 1137.871371] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.880547] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464997, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.909261] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464996, 'name': Rename_Task, 'duration_secs': 0.135248} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.909551] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1137.910063] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2865da52-1d96-4e0e-af87-204998282eef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.912200] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.912437] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.912662] env[61728]: DEBUG nova.objects.instance [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Lazy-loading 'resources' on Instance uuid dbe877e6-1754-4cf7-b454-abd88bfa713a {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1137.919179] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1137.919179] env[61728]: value = "task-464998" [ 1137.919179] env[61728]: _type = "Task" [ 1137.919179] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.923208] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.933073] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464998, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.002641] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528cc1a8-6651-18c0-a1d7-fade2928b28f, 'name': SearchDatastore_Task, 'duration_secs': 0.010918} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.003522] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2915e3de-b144-45e8-b8a1-90fac2783840 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.009199] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1138.009199] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52abc668-4dfb-33ef-6e6b-52c118d0dffa" [ 1138.009199] env[61728]: _type = "Task" [ 1138.009199] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.019919] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52abc668-4dfb-33ef-6e6b-52c118d0dffa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.240864] env[61728]: DEBUG nova.compute.manager [req-35f36308-602d-4154-9e2c-568c54d9c21b req-ee36b834-5c10-45ae-b07f-c50d7bbf17b5 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Received event network-changed-d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1138.241214] env[61728]: DEBUG nova.compute.manager [req-35f36308-602d-4154-9e2c-568c54d9c21b req-ee36b834-5c10-45ae-b07f-c50d7bbf17b5 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Refreshing instance network info cache due to event network-changed-d04bc3a5-d378-4e1e-a145-ec523bcc442a. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1138.241416] env[61728]: DEBUG oslo_concurrency.lockutils [req-35f36308-602d-4154-9e2c-568c54d9c21b req-ee36b834-5c10-45ae-b07f-c50d7bbf17b5 service nova] Acquiring lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1138.241565] env[61728]: DEBUG oslo_concurrency.lockutils [req-35f36308-602d-4154-9e2c-568c54d9c21b req-ee36b834-5c10-45ae-b07f-c50d7bbf17b5 service nova] Acquired lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.241730] env[61728]: DEBUG nova.network.neutron [req-35f36308-602d-4154-9e2c-568c54d9c21b req-ee36b834-5c10-45ae-b07f-c50d7bbf17b5 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Refreshing network info cache for port d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1138.316530] env[61728]: DEBUG oslo_vmware.api [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-464990, 'name': ReconfigVM_Task, 'duration_secs': 1.788994} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.316830] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122212', 'volume_id': '0136be7a-12a6-43a4-847f-203c70cf8dc8', 'name': 'volume-0136be7a-12a6-43a4-847f-203c70cf8dc8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd55beb02-00fc-4df6-a239-9e5d776a7c18', 'attached_at': '', 'detached_at': '', 'volume_id': '0136be7a-12a6-43a4-847f-203c70cf8dc8', 'serial': '0136be7a-12a6-43a4-847f-203c70cf8dc8'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1138.367016] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.367336] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.367555] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "8c5b2380-bd14-451e-9612-2767c2a6adb2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.367739] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.367915] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.370081] env[61728]: INFO nova.compute.manager [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Terminating instance [ 1138.372028] env[61728]: DEBUG nova.compute.manager [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1138.372237] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1138.373182] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1a1054f-ab7e-4f86-b672-88535d643170 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.384379] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464997, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.386353] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1138.386595] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a1d67e27-af19-487a-b53a-ff89e3d73892 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.392803] env[61728]: DEBUG oslo_vmware.api [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1138.392803] env[61728]: value = "task-464999" [ 1138.392803] env[61728]: _type = "Task" [ 1138.392803] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.400329] env[61728]: DEBUG oslo_vmware.api [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464999, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.429702] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464998, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.524127] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52abc668-4dfb-33ef-6e6b-52c118d0dffa, 'name': SearchDatastore_Task, 'duration_secs': 0.011456} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.524440] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1138.524742] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] c92e595d-c686-4810-bb39-d98de4141121/c92e595d-c686-4810-bb39-d98de4141121.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1138.525109] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ea56c11-00da-485c-983e-55bcdc7da75f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.535548] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1138.535548] env[61728]: value = "task-465000" [ 1138.535548] env[61728]: _type = "Task" [ 1138.535548] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.546155] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465000, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.556016] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fdf9dc5-0bf6-4e06-8812-e92fafb5a0ea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.565159] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8abb0453-e235-4f0f-b288-501640adb117 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.595686] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a735d1c7-f82d-4d4a-93fb-6a187179677b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.603150] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-640399fe-b804-4c8a-a0f7-93a3a3582bce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.616029] env[61728]: DEBUG nova.compute.provider_tree [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1138.769560] env[61728]: DEBUG nova.objects.instance [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lazy-loading 'flavor' on Instance uuid c1ed0e8b-5cf9-4ff0-931e-fadc3432a125 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1138.858962] env[61728]: DEBUG nova.objects.instance [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lazy-loading 'flavor' on Instance uuid d55beb02-00fc-4df6-a239-9e5d776a7c18 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1138.882768] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464997, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.904638] env[61728]: DEBUG oslo_vmware.api [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-464999, 'name': PowerOffVM_Task, 'duration_secs': 0.192696} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.905586] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1138.905586] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1138.905586] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e61bdeec-a70a-42b4-8db0-c75dea34f321 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.929196] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464998, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.971445] env[61728]: DEBUG nova.network.neutron [req-35f36308-602d-4154-9e2c-568c54d9c21b req-ee36b834-5c10-45ae-b07f-c50d7bbf17b5 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updated VIF entry in instance network info cache for port d04bc3a5-d378-4e1e-a145-ec523bcc442a. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1138.971982] env[61728]: DEBUG nova.network.neutron [req-35f36308-602d-4154-9e2c-568c54d9c21b req-ee36b834-5c10-45ae-b07f-c50d7bbf17b5 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updating instance_info_cache with network_info: [{"id": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "address": "fa:16:3e:f1:65:32", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapd04bc3a5-d3", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.996486] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "532507fa-996d-45c7-bf70-f09c93be79ed" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.045729] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465000, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.119408] env[61728]: DEBUG nova.scheduler.client.report [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1139.276155] env[61728]: DEBUG oslo_concurrency.lockutils [None req-12ba502b-a7ac-46fe-b411-4eb8ee2867f6 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.301s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.326384] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1139.326658] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1139.326984] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleting the datastore file [datastore1] 8c5b2380-bd14-451e-9612-2767c2a6adb2 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1139.327825] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bc6f4d18-99a9-4fc6-82f4-30bb51c40532 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.335012] env[61728]: DEBUG oslo_vmware.api [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1139.335012] env[61728]: value = "task-465002" [ 1139.335012] env[61728]: _type = "Task" [ 1139.335012] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.342851] env[61728]: DEBUG oslo_vmware.api [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465002, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.383312] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-464997, 'name': ReconfigVM_Task, 'duration_secs': 1.380312} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.383635] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Reconfigured VM instance instance-00000076 to attach disk [datastore1] 2db96467-5fb4-49f4-954b-dbdd21e46232/2db96467-5fb4-49f4-954b-dbdd21e46232.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1139.384329] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d07c32c1-e630-47b0-a778-530db18c8af9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.389891] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1139.389891] env[61728]: value = "task-465003" [ 1139.389891] env[61728]: _type = "Task" [ 1139.389891] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.400017] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465003, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.429226] env[61728]: DEBUG oslo_vmware.api [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-464998, 'name': PowerOnVM_Task, 'duration_secs': 1.500207} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.429505] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1139.475315] env[61728]: DEBUG oslo_concurrency.lockutils [req-35f36308-602d-4154-9e2c-568c54d9c21b req-ee36b834-5c10-45ae-b07f-c50d7bbf17b5 service nova] Releasing lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.487211] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.487490] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.536344] env[61728]: DEBUG nova.compute.manager [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1139.537470] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba27eb63-b81a-4422-a37f-ffad5703a137 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.554542] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465000, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.673681} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.555021] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] c92e595d-c686-4810-bb39-d98de4141121/c92e595d-c686-4810-bb39-d98de4141121.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1139.555281] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1139.555535] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8f664800-12b8-4652-8e07-f99dfdf902c2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.560807] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1139.560807] env[61728]: value = "task-465004" [ 1139.560807] env[61728]: _type = "Task" [ 1139.560807] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.568592] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465004, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.625464] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.713s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.627821] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.705s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.628076] env[61728]: DEBUG nova.objects.instance [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lazy-loading 'resources' on Instance uuid 532507fa-996d-45c7-bf70-f09c93be79ed {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1139.647608] env[61728]: INFO nova.scheduler.client.report [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Deleted allocations for instance dbe877e6-1754-4cf7-b454-abd88bfa713a [ 1139.846879] env[61728]: DEBUG oslo_vmware.api [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465002, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204439} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.847211] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1139.847430] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1139.847850] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1139.847850] env[61728]: INFO nova.compute.manager [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Took 1.48 seconds to destroy the instance on the hypervisor. [ 1139.848038] env[61728]: DEBUG oslo.service.loopingcall [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1139.848241] env[61728]: DEBUG nova.compute.manager [-] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1139.848338] env[61728]: DEBUG nova.network.neutron [-] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1139.866550] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2170e30c-8907-4555-93d4-a6d35b04173a tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.771s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.899484] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465003, 'name': Rename_Task, 'duration_secs': 0.166112} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.899484] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1139.899773] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a2dfdb5e-3a5c-4373-83d6-5771a25f105e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.907867] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1139.907867] env[61728]: value = "task-465005" [ 1139.907867] env[61728]: _type = "Task" [ 1139.907867] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.917093] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465005, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.990775] env[61728]: INFO nova.compute.manager [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Detaching volume 4b08464d-0026-4ba9-a28e-88e941c8f364 [ 1140.032849] env[61728]: INFO nova.virt.block_device [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Attempting to driver detach volume 4b08464d-0026-4ba9-a28e-88e941c8f364 from mountpoint /dev/sdb [ 1140.033208] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1140.033474] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122229', 'volume_id': '4b08464d-0026-4ba9-a28e-88e941c8f364', 'name': 'volume-4b08464d-0026-4ba9-a28e-88e941c8f364', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c1ed0e8b-5cf9-4ff0-931e-fadc3432a125', 'attached_at': '', 'detached_at': '', 'volume_id': '4b08464d-0026-4ba9-a28e-88e941c8f364', 'serial': '4b08464d-0026-4ba9-a28e-88e941c8f364'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1140.034468] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfede51f-230b-4ff9-b266-fe9036366213 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.060534] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bad9e0b-16e3-4a21-98a2-e856cd8cd2e1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.068930] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f72961de-5055-45d5-85ad-3aabcc240a36 tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 23.925s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.078459] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2747b8a1-2068-4501-a328-8593c386cd94 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.081125] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465004, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061809} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.081435] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1140.082520] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dfad8d7-916f-42b3-ab4f-d86fc471ceec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.101770] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9072790-431f-4d65-b181-7bef21e98d8a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.122652] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Reconfiguring VM instance instance-00000077 to attach disk [datastore1] c92e595d-c686-4810-bb39-d98de4141121/c92e595d-c686-4810-bb39-d98de4141121.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1140.123617] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab30b4c2-ef04-407e-a505-cd05931aa207 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.151911] env[61728]: DEBUG nova.objects.instance [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lazy-loading 'numa_topology' on Instance uuid 532507fa-996d-45c7-bf70-f09c93be79ed {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.157018] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] The volume has not been displaced from its original location: [datastore1] volume-4b08464d-0026-4ba9-a28e-88e941c8f364/volume-4b08464d-0026-4ba9-a28e-88e941c8f364.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1140.159449] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Reconfiguring VM instance instance-00000072 to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1140.163029] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dee14889-afea-459f-91f2-85d6fd6c927d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.177172] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3d59fa6a-1de1-4077-80cf-c4fc768c4e23 tempest-ServerShowV257Test-637142476 tempest-ServerShowV257Test-637142476-project-member] Lock "dbe877e6-1754-4cf7-b454-abd88bfa713a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.981s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.178477] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1140.178477] env[61728]: value = "task-465006" [ 1140.178477] env[61728]: _type = "Task" [ 1140.178477] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.183825] env[61728]: DEBUG oslo_vmware.api [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1140.183825] env[61728]: value = "task-465007" [ 1140.183825] env[61728]: _type = "Task" [ 1140.183825] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.191062] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465006, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.195954] env[61728]: DEBUG oslo_vmware.api [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-465007, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.307560] env[61728]: DEBUG nova.compute.manager [req-f6282e32-e317-4ac3-b6fd-0e921e61431e req-221c2128-b6ef-4d26-ab65-a9cfec17e9b6 service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Received event network-vif-deleted-4660e4da-30e8-4e51-9aea-a9375732537f {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1140.307650] env[61728]: INFO nova.compute.manager [req-f6282e32-e317-4ac3-b6fd-0e921e61431e req-221c2128-b6ef-4d26-ab65-a9cfec17e9b6 service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Neutron deleted interface 4660e4da-30e8-4e51-9aea-a9375732537f; detaching it from the instance and deleting it from the info cache [ 1140.307823] env[61728]: DEBUG nova.network.neutron [req-f6282e32-e317-4ac3-b6fd-0e921e61431e req-221c2128-b6ef-4d26-ab65-a9cfec17e9b6 service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.330244] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0340b421-2fbc-4a52-9674-6c03396c44b0 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.330529] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0340b421-2fbc-4a52-9674-6c03396c44b0 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.330740] env[61728]: DEBUG nova.compute.manager [None req-0340b421-2fbc-4a52-9674-6c03396c44b0 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1140.331640] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6656724f-1741-4e00-acc2-4c5a736fc002 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.338607] env[61728]: DEBUG nova.compute.manager [None req-0340b421-2fbc-4a52-9674-6c03396c44b0 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61728) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1140.339166] env[61728]: DEBUG nova.objects.instance [None req-0340b421-2fbc-4a52-9674-6c03396c44b0 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lazy-loading 'flavor' on Instance uuid d55beb02-00fc-4df6-a239-9e5d776a7c18 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.417809] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465005, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.661233] env[61728]: DEBUG nova.objects.base [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Object Instance<532507fa-996d-45c7-bf70-f09c93be79ed> lazy-loaded attributes: resources,numa_topology {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1140.695988] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465006, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.701385] env[61728]: DEBUG oslo_vmware.api [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-465007, 'name': ReconfigVM_Task, 'duration_secs': 0.265741} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.701844] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Reconfigured VM instance instance-00000072 to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1140.709705] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf3346b8-9d26-4d2d-aec8-37ff66f8e168 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.726465] env[61728]: DEBUG oslo_vmware.api [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1140.726465] env[61728]: value = "task-465008" [ 1140.726465] env[61728]: _type = "Task" [ 1140.726465] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.735652] env[61728]: DEBUG oslo_vmware.api [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-465008, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.779935] env[61728]: DEBUG nova.network.neutron [-] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.815225] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e4523a1-921d-4388-b69d-22bd867b4702 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.827905] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab4573b-2eac-42da-a4af-fe3917cd155b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.849343] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0340b421-2fbc-4a52-9674-6c03396c44b0 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1140.849652] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d5fc863-940a-4d1f-bb77-8b3209b360ec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.863249] env[61728]: DEBUG oslo_vmware.api [None req-0340b421-2fbc-4a52-9674-6c03396c44b0 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1140.863249] env[61728]: value = "task-465009" [ 1140.863249] env[61728]: _type = "Task" [ 1140.863249] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.875259] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4ec7d7-6f94-4b48-9370-666017ac3fde {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.878740] env[61728]: DEBUG nova.compute.manager [req-f6282e32-e317-4ac3-b6fd-0e921e61431e req-221c2128-b6ef-4d26-ab65-a9cfec17e9b6 service nova] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Detach interface failed, port_id=4660e4da-30e8-4e51-9aea-a9375732537f, reason: Instance 8c5b2380-bd14-451e-9612-2767c2a6adb2 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1140.893027] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a7e803-880e-4df5-b5de-c854eb1ef96b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.895635] env[61728]: DEBUG oslo_vmware.api [None req-0340b421-2fbc-4a52-9674-6c03396c44b0 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465009, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.927349] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69cbf8b3-7bd6-4d47-b7e3-f1b49baac80a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.935125] env[61728]: DEBUG oslo_vmware.api [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465005, 'name': PowerOnVM_Task, 'duration_secs': 0.978536} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.937305] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1140.937718] env[61728]: INFO nova.compute.manager [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Took 11.93 seconds to spawn the instance on the hypervisor. [ 1140.937718] env[61728]: DEBUG nova.compute.manager [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1140.938536] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ef0704-2954-46c3-bc5c-5bd6b13da4a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.941908] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e51c837-1417-410f-961b-e6f7a51b422a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.956036] env[61728]: DEBUG nova.compute.provider_tree [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.191663] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465006, 'name': ReconfigVM_Task, 'duration_secs': 0.688941} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.191979] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Reconfigured VM instance instance-00000077 to attach disk [datastore1] c92e595d-c686-4810-bb39-d98de4141121/c92e595d-c686-4810-bb39-d98de4141121.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1141.192604] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e9dd076-9f88-484c-8814-0e8ae79c8369 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.198585] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1141.198585] env[61728]: value = "task-465010" [ 1141.198585] env[61728]: _type = "Task" [ 1141.198585] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.206103] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465010, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.235548] env[61728]: DEBUG oslo_vmware.api [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-465008, 'name': ReconfigVM_Task, 'duration_secs': 0.156926} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.235869] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122229', 'volume_id': '4b08464d-0026-4ba9-a28e-88e941c8f364', 'name': 'volume-4b08464d-0026-4ba9-a28e-88e941c8f364', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c1ed0e8b-5cf9-4ff0-931e-fadc3432a125', 'attached_at': '', 'detached_at': '', 'volume_id': '4b08464d-0026-4ba9-a28e-88e941c8f364', 'serial': '4b08464d-0026-4ba9-a28e-88e941c8f364'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1141.282662] env[61728]: INFO nova.compute.manager [-] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Took 1.43 seconds to deallocate network for instance. [ 1141.384635] env[61728]: DEBUG oslo_vmware.api [None req-0340b421-2fbc-4a52-9674-6c03396c44b0 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465009, 'name': PowerOffVM_Task, 'duration_secs': 0.221622} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.384917] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0340b421-2fbc-4a52-9674-6c03396c44b0 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1141.385114] env[61728]: DEBUG nova.compute.manager [None req-0340b421-2fbc-4a52-9674-6c03396c44b0 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1141.385843] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887c0872-3b3d-48b4-9388-ea071fa50719 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.462079] env[61728]: DEBUG nova.scheduler.client.report [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1141.471416] env[61728]: INFO nova.compute.manager [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Took 20.86 seconds to build instance. [ 1141.497875] env[61728]: DEBUG oslo_concurrency.lockutils [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.498132] env[61728]: DEBUG oslo_concurrency.lockutils [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.498440] env[61728]: DEBUG oslo_concurrency.lockutils [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.498638] env[61728]: DEBUG oslo_concurrency.lockutils [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.498812] env[61728]: DEBUG oslo_concurrency.lockutils [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.500697] env[61728]: INFO nova.compute.manager [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Terminating instance [ 1141.502471] env[61728]: DEBUG nova.compute.manager [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1141.502675] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1141.503706] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef043c4c-a40d-459a-8c37-98c0cefe6e1f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.510984] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1141.511223] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ab99057a-6d8e-4905-bc3d-8d471279ead2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.517633] env[61728]: DEBUG oslo_vmware.api [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1141.517633] env[61728]: value = "task-465011" [ 1141.517633] env[61728]: _type = "Task" [ 1141.517633] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.526569] env[61728]: DEBUG oslo_vmware.api [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-465011, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.709377] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465010, 'name': Rename_Task, 'duration_secs': 0.176952} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.709759] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1141.710073] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-384b5715-a7e0-4a03-b506-3776ea84a44f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.716383] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1141.716383] env[61728]: value = "task-465012" [ 1141.716383] env[61728]: _type = "Task" [ 1141.716383] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.724012] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465012, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.784007] env[61728]: DEBUG nova.objects.instance [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lazy-loading 'flavor' on Instance uuid c1ed0e8b-5cf9-4ff0-931e-fadc3432a125 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1141.789062] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.897527] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0340b421-2fbc-4a52-9674-6c03396c44b0 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.567s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.970079] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.342s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.973320] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.184s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.973895] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.975809] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b3894512-5fd5-4323-b3f6-694fe2c76fab tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "2db96467-5fb4-49f4-954b-dbdd21e46232" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.375s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.995813] env[61728]: INFO nova.scheduler.client.report [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleted allocations for instance 8c5b2380-bd14-451e-9612-2767c2a6adb2 [ 1142.030622] env[61728]: DEBUG oslo_vmware.api [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-465011, 'name': PowerOffVM_Task, 'duration_secs': 0.250159} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.030840] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1142.031130] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1142.031484] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2298748a-cedd-480a-8331-acc2b4500883 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.099085] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1142.099330] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1142.099520] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleting the datastore file [datastore1] 77b8f509-17e7-4c25-a43d-1655c16ce8d1 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1142.099839] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-374b1a38-d8c6-4781-b8f3-341592be4418 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.105803] env[61728]: DEBUG oslo_vmware.api [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for the task: (returnval){ [ 1142.105803] env[61728]: value = "task-465014" [ 1142.105803] env[61728]: _type = "Task" [ 1142.105803] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.115121] env[61728]: DEBUG oslo_vmware.api [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-465014, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.227473] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465012, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.477876] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bdf4f3a-66e9-4603-8807-3b350311fe79 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "532507fa-996d-45c7-bf70-f09c93be79ed" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.593s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.479072] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "532507fa-996d-45c7-bf70-f09c93be79ed" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 3.483s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.479217] env[61728]: INFO nova.compute.manager [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Unshelving [ 1142.503929] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1e4dbdb1-0545-42ce-b92c-b14b6e14d216 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "8c5b2380-bd14-451e-9612-2767c2a6adb2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.136s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.558461] env[61728]: DEBUG nova.compute.manager [req-137fc50f-ad2c-464e-abbe-62c80669e578 req-2e7a45e4-8a0f-48e5-b45c-e0af7d5d5c27 service nova] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Received event network-changed-e6562f3d-358f-482b-b0a1-b6a09c232c71 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1142.558591] env[61728]: DEBUG nova.compute.manager [req-137fc50f-ad2c-464e-abbe-62c80669e578 req-2e7a45e4-8a0f-48e5-b45c-e0af7d5d5c27 service nova] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Refreshing instance network info cache due to event network-changed-e6562f3d-358f-482b-b0a1-b6a09c232c71. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1142.558807] env[61728]: DEBUG oslo_concurrency.lockutils [req-137fc50f-ad2c-464e-abbe-62c80669e578 req-2e7a45e4-8a0f-48e5-b45c-e0af7d5d5c27 service nova] Acquiring lock "refresh_cache-2db96467-5fb4-49f4-954b-dbdd21e46232" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.558954] env[61728]: DEBUG oslo_concurrency.lockutils [req-137fc50f-ad2c-464e-abbe-62c80669e578 req-2e7a45e4-8a0f-48e5-b45c-e0af7d5d5c27 service nova] Acquired lock "refresh_cache-2db96467-5fb4-49f4-954b-dbdd21e46232" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.561014] env[61728]: DEBUG nova.network.neutron [req-137fc50f-ad2c-464e-abbe-62c80669e578 req-2e7a45e4-8a0f-48e5-b45c-e0af7d5d5c27 service nova] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Refreshing network info cache for port e6562f3d-358f-482b-b0a1-b6a09c232c71 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1142.619271] env[61728]: DEBUG oslo_vmware.api [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Task: {'id': task-465014, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20219} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.619703] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1142.619977] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1142.620183] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1142.620362] env[61728]: INFO nova.compute.manager [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1142.620601] env[61728]: DEBUG oslo.service.loopingcall [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1142.620791] env[61728]: DEBUG nova.compute.manager [-] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1142.620888] env[61728]: DEBUG nova.network.neutron [-] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1142.727559] env[61728]: DEBUG oslo_vmware.api [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465012, 'name': PowerOnVM_Task, 'duration_secs': 0.720947} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.728092] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1142.728406] env[61728]: INFO nova.compute.manager [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Took 11.40 seconds to spawn the instance on the hypervisor. [ 1142.728655] env[61728]: DEBUG nova.compute.manager [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1142.729508] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dfbb196-3263-4eb3-8bbe-04e2579c1f0a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.792562] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9a64861b-8c43-4e99-8d87-3ce80a0ca6b9 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.305s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.965878] env[61728]: DEBUG nova.objects.instance [None req-2cc073a9-9741-4ed4-9ab7-d007d54bcd67 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lazy-loading 'flavor' on Instance uuid d55beb02-00fc-4df6-a239-9e5d776a7c18 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1143.123107] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.123107] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.123107] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.123107] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.123107] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.124112] env[61728]: INFO nova.compute.manager [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Terminating instance [ 1143.126109] env[61728]: DEBUG nova.compute.manager [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1143.126468] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1143.129014] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ed3fb2-a508-4e51-a557-508c0959f914 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.138153] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1143.138609] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3aeffed7-c040-4d50-9dc7-964b99490a64 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.148016] env[61728]: DEBUG oslo_vmware.api [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1143.148016] env[61728]: value = "task-465015" [ 1143.148016] env[61728]: _type = "Task" [ 1143.148016] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.158057] env[61728]: DEBUG oslo_vmware.api [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-465015, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.234253] env[61728]: DEBUG nova.compute.manager [req-599015bb-d97a-4f26-85b5-c4a405cd5ed8 req-6c9b1253-8dae-4cd3-896b-f163e27fc6a7 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Received event network-vif-deleted-ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1143.234528] env[61728]: INFO nova.compute.manager [req-599015bb-d97a-4f26-85b5-c4a405cd5ed8 req-6c9b1253-8dae-4cd3-896b-f163e27fc6a7 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Neutron deleted interface ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb; detaching it from the instance and deleting it from the info cache [ 1143.234833] env[61728]: DEBUG nova.network.neutron [req-599015bb-d97a-4f26-85b5-c4a405cd5ed8 req-6c9b1253-8dae-4cd3-896b-f163e27fc6a7 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.246712] env[61728]: INFO nova.compute.manager [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Took 16.22 seconds to build instance. [ 1143.293909] env[61728]: DEBUG nova.network.neutron [req-137fc50f-ad2c-464e-abbe-62c80669e578 req-2e7a45e4-8a0f-48e5-b45c-e0af7d5d5c27 service nova] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Updated VIF entry in instance network info cache for port e6562f3d-358f-482b-b0a1-b6a09c232c71. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1143.294347] env[61728]: DEBUG nova.network.neutron [req-137fc50f-ad2c-464e-abbe-62c80669e578 req-2e7a45e4-8a0f-48e5-b45c-e0af7d5d5c27 service nova] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Updating instance_info_cache with network_info: [{"id": "e6562f3d-358f-482b-b0a1-b6a09c232c71", "address": "fa:16:3e:8a:36:db", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6562f3d-35", "ovs_interfaceid": "e6562f3d-358f-482b-b0a1-b6a09c232c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.474096] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cc073a9-9741-4ed4-9ab7-d007d54bcd67 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1143.474291] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cc073a9-9741-4ed4-9ab7-d007d54bcd67 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquired lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.474484] env[61728]: DEBUG nova.network.neutron [None req-2cc073a9-9741-4ed4-9ab7-d007d54bcd67 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1143.474693] env[61728]: DEBUG nova.objects.instance [None req-2cc073a9-9741-4ed4-9ab7-d007d54bcd67 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lazy-loading 'info_cache' on Instance uuid d55beb02-00fc-4df6-a239-9e5d776a7c18 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1143.487504] env[61728]: DEBUG nova.compute.utils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1143.530650] env[61728]: DEBUG nova.network.neutron [-] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.656372] env[61728]: DEBUG oslo_vmware.api [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-465015, 'name': PowerOffVM_Task, 'duration_secs': 0.18342} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.656715] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1143.656909] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1143.657190] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0c24e9e5-e9c6-4c57-bb3e-985aa028ad8d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.714584] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1143.714832] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1143.715044] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Deleting the datastore file [datastore1] c1ed0e8b-5cf9-4ff0-931e-fadc3432a125 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1143.715333] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7128a1a8-e96e-48fe-bdda-da092ec989fe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.721488] env[61728]: DEBUG oslo_vmware.api [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for the task: (returnval){ [ 1143.721488] env[61728]: value = "task-465017" [ 1143.721488] env[61728]: _type = "Task" [ 1143.721488] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.729377] env[61728]: DEBUG oslo_vmware.api [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-465017, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.737083] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c4b9f7c0-6e67-470e-989a-d04ab0bd7bf3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.746031] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0585767d-c0c5-43d3-bc5f-6965aef7b61b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.757243] env[61728]: DEBUG oslo_concurrency.lockutils [None req-deadfa6f-e77c-4a6a-96a3-362462432c01 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "c92e595d-c686-4810-bb39-d98de4141121" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.737s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.775639] env[61728]: DEBUG nova.compute.manager [req-599015bb-d97a-4f26-85b5-c4a405cd5ed8 req-6c9b1253-8dae-4cd3-896b-f163e27fc6a7 service nova] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Detach interface failed, port_id=ea6ae87d-d3a2-4ecb-b14e-b3e0f99e11eb, reason: Instance 77b8f509-17e7-4c25-a43d-1655c16ce8d1 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1143.796666] env[61728]: DEBUG oslo_concurrency.lockutils [req-137fc50f-ad2c-464e-abbe-62c80669e578 req-2e7a45e4-8a0f-48e5-b45c-e0af7d5d5c27 service nova] Releasing lock "refresh_cache-2db96467-5fb4-49f4-954b-dbdd21e46232" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.818236] env[61728]: INFO nova.compute.manager [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Rescuing [ 1143.818686] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "refresh_cache-c92e595d-c686-4810-bb39-d98de4141121" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1143.819397] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquired lock "refresh_cache-c92e595d-c686-4810-bb39-d98de4141121" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.819397] env[61728]: DEBUG nova.network.neutron [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1143.876378] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.876617] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.978833] env[61728]: DEBUG nova.objects.base [None req-2cc073a9-9741-4ed4-9ab7-d007d54bcd67 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1143.991483] env[61728]: INFO nova.virt.block_device [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Booting with volume ad2189db-dfcc-42fb-ab2a-15897b9ab1c8 at /dev/sdb [ 1144.031146] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c26d43c-638d-4f0c-90dd-5d3784332526 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.033927] env[61728]: INFO nova.compute.manager [-] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Took 1.41 seconds to deallocate network for instance. [ 1144.044638] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567f5269-0382-471c-bf2e-7b7023f6a866 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.075554] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d452474b-4379-44d2-a5fb-e6b425c70615 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.083940] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc7d35b-d407-40a5-90eb-c22081bb9229 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.110329] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3591665-a8f6-4690-9972-9ec5762b5b24 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.116466] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7098f2-312e-4e3d-a1b2-b2e9a5184baa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.128912] env[61728]: DEBUG nova.virt.block_device [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updating existing volume attachment record: c4b63838-2ce5-4b66-b34a-7bcd8dc37430 {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1144.232542] env[61728]: DEBUG oslo_vmware.api [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Task: {'id': task-465017, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.324797} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.232819] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1144.233059] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1144.233287] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1144.233507] env[61728]: INFO nova.compute.manager [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1144.233805] env[61728]: DEBUG oslo.service.loopingcall [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1144.234051] env[61728]: DEBUG nova.compute.manager [-] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1144.234157] env[61728]: DEBUG nova.network.neutron [-] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1144.379301] env[61728]: DEBUG nova.compute.manager [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1144.542941] env[61728]: DEBUG oslo_concurrency.lockutils [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.543376] env[61728]: DEBUG oslo_concurrency.lockutils [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.543544] env[61728]: DEBUG nova.objects.instance [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lazy-loading 'resources' on Instance uuid 77b8f509-17e7-4c25-a43d-1655c16ce8d1 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1144.635971] env[61728]: DEBUG nova.network.neutron [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Updating instance_info_cache with network_info: [{"id": "7aee91de-db46-4071-8f9e-980bcfc29365", "address": "fa:16:3e:65:1f:da", "network": {"id": "4bb51519-042f-44ea-9325-a479be9790d3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1499510217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "909c86730f554a04ab05dcad3a53e9dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7aee91de-db", "ovs_interfaceid": "7aee91de-db46-4071-8f9e-980bcfc29365", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.767581] env[61728]: DEBUG nova.network.neutron [None req-2cc073a9-9741-4ed4-9ab7-d007d54bcd67 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Updating instance_info_cache with network_info: [{"id": "4155b1a7-975f-404c-8374-47665c8046ca", "address": "fa:16:3e:e4:b3:9b", "network": {"id": "72d11e2a-907e-419e-b8a4-53aa39f2566a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1755247352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833a646ea21a48c4951ed8cedebaa98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3836fb52-19c6-4e10-a0ca-f0bca73dc887", "external-id": "nsx-vlan-transportzone-964", "segmentation_id": 964, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4155b1a7-97", "ovs_interfaceid": "4155b1a7-975f-404c-8374-47665c8046ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.916904] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.138766] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Releasing lock "refresh_cache-c92e595d-c686-4810-bb39-d98de4141121" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.169167] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6dfb7f5-1630-4f4d-b346-4d70b19c74e7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.178118] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b055e0d-341a-489c-9918-856174d7906f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.208712] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f709a2-37df-494c-95e6-107b84a96e3c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.216019] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afdcc623-4de1-48c7-b635-8084c3318d31 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.229885] env[61728]: DEBUG nova.compute.provider_tree [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1145.260022] env[61728]: DEBUG nova.compute.manager [req-da2d6c3b-7714-4900-9b74-cbbb1f2b34ec req-aa0af030-bd89-4ec3-94f4-1264c700c882 service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Received event network-vif-deleted-4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1145.260175] env[61728]: INFO nova.compute.manager [req-da2d6c3b-7714-4900-9b74-cbbb1f2b34ec req-aa0af030-bd89-4ec3-94f4-1264c700c882 service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Neutron deleted interface 4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd; detaching it from the instance and deleting it from the info cache [ 1145.260343] env[61728]: DEBUG nova.network.neutron [req-da2d6c3b-7714-4900-9b74-cbbb1f2b34ec req-aa0af030-bd89-4ec3-94f4-1264c700c882 service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.269970] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2cc073a9-9741-4ed4-9ab7-d007d54bcd67 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Releasing lock "refresh_cache-d55beb02-00fc-4df6-a239-9e5d776a7c18" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.292815] env[61728]: DEBUG nova.network.neutron [-] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.675609] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1145.675773] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b05d0969-12eb-408c-9ded-999aae3a5c24 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.683794] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1145.683794] env[61728]: value = "task-465022" [ 1145.683794] env[61728]: _type = "Task" [ 1145.683794] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.692516] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465022, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.732701] env[61728]: DEBUG nova.scheduler.client.report [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1145.762420] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6288a138-11a8-490f-86ae-3288e3d7043b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.771986] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57596c86-08bf-47cf-9fb9-86deab32cc16 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.783202] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cc073a9-9741-4ed4-9ab7-d007d54bcd67 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1145.783862] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9dfb3798-09a2-498b-8c0d-296e0215663b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.802463] env[61728]: INFO nova.compute.manager [-] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Took 1.57 seconds to deallocate network for instance. [ 1145.802779] env[61728]: DEBUG nova.compute.manager [req-da2d6c3b-7714-4900-9b74-cbbb1f2b34ec req-aa0af030-bd89-4ec3-94f4-1264c700c882 service nova] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Detach interface failed, port_id=4c25f3d2-3ec3-4a7e-b8f8-347d7269aedd, reason: Instance c1ed0e8b-5cf9-4ff0-931e-fadc3432a125 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1145.805661] env[61728]: DEBUG oslo_vmware.api [None req-2cc073a9-9741-4ed4-9ab7-d007d54bcd67 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1145.805661] env[61728]: value = "task-465023" [ 1145.805661] env[61728]: _type = "Task" [ 1145.805661] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.816322] env[61728]: DEBUG oslo_vmware.api [None req-2cc073a9-9741-4ed4-9ab7-d007d54bcd67 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465023, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.193466] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465022, 'name': PowerOffVM_Task, 'duration_secs': 0.349739} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.193759] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1146.194530] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29048d83-4b75-4012-a19f-854857390c44 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.211673] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468abac2-ed1e-4517-b2f9-699d4d42b64a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.238747] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1146.239573] env[61728]: DEBUG oslo_concurrency.lockutils [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.696s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.241559] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc2e2b9f-b109-4a5e-bab1-91d0ea8ccc19 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.243403] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.327s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.245180] env[61728]: INFO nova.compute.claims [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1146.253214] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1146.253214] env[61728]: value = "task-465024" [ 1146.253214] env[61728]: _type = "Task" [ 1146.253214] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.261450] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1146.261665] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1146.261976] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.262203] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.262435] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1146.262609] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6639f0d1-c905-4ada-a281-76dafc8d9b2d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.266808] env[61728]: INFO nova.scheduler.client.report [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Deleted allocations for instance 77b8f509-17e7-4c25-a43d-1655c16ce8d1 [ 1146.276661] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1146.276771] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1146.277901] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97b46549-cad3-474b-9120-64940c062a0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.282750] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1146.282750] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d683b0-74da-fb54-7497-9a74774e874e" [ 1146.282750] env[61728]: _type = "Task" [ 1146.282750] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.291037] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d683b0-74da-fb54-7497-9a74774e874e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.314523] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.319443] env[61728]: DEBUG oslo_vmware.api [None req-2cc073a9-9741-4ed4-9ab7-d007d54bcd67 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465023, 'name': PowerOnVM_Task, 'duration_secs': 0.381207} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.319755] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cc073a9-9741-4ed4-9ab7-d007d54bcd67 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1146.319945] env[61728]: DEBUG nova.compute.manager [None req-2cc073a9-9741-4ed4-9ab7-d007d54bcd67 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1146.320701] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-180b5ae2-bcbc-4f77-a4e9-fe5a0f581cf3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.775059] env[61728]: DEBUG oslo_concurrency.lockutils [None req-665555b5-bcb9-4620-b8cb-e8dc7d82cc6e tempest-ServerActionsTestOtherB-564458324 tempest-ServerActionsTestOtherB-564458324-project-member] Lock "77b8f509-17e7-4c25-a43d-1655c16ce8d1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.277s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.794019] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52d683b0-74da-fb54-7497-9a74774e874e, 'name': SearchDatastore_Task, 'duration_secs': 0.023063} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.794857] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c3e35e5-3f6e-4e6a-8824-cee5a63b09f6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.801883] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1146.801883] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52351275-710d-2b4c-2064-0553dbfc924c" [ 1146.801883] env[61728]: _type = "Task" [ 1146.801883] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.811782] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52351275-710d-2b4c-2064-0553dbfc924c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.314576] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52351275-710d-2b4c-2064-0553dbfc924c, 'name': SearchDatastore_Task, 'duration_secs': 0.062542} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.314867] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.315120] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] c92e595d-c686-4810-bb39-d98de4141121/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk. {{(pid=61728) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1147.315385] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e590d95c-a63c-4d69-81fc-ffcea779339a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.324465] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1147.324465] env[61728]: value = "task-465028" [ 1147.324465] env[61728]: _type = "Task" [ 1147.324465] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.334016] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465028, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.358360] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0a713e-b199-4bdb-bfd7-f234b39228dd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.365668] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff539a5-ff81-4d4c-8738-93b9b4147eee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.396205] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d86ce7-a22c-48a9-8393-a32333a6698e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.403789] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4fb1141-dba8-4f19-a422-def2b4808527 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.416327] env[61728]: DEBUG nova.compute.provider_tree [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1147.834631] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465028, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.938068] env[61728]: ERROR nova.scheduler.client.report [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [req-ef3d1576-66be-4422-b06f-37c399265f89] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ef3d1576-66be-4422-b06f-37c399265f89"}]} [ 1147.954611] env[61728]: DEBUG nova.scheduler.client.report [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1147.967858] env[61728]: DEBUG nova.scheduler.client.report [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1147.968123] env[61728]: DEBUG nova.compute.provider_tree [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1147.980149] env[61728]: DEBUG nova.scheduler.client.report [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1148.003598] env[61728]: DEBUG nova.scheduler.client.report [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1148.118619] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db59005e-1a51-48a1-8f49-930277ed321b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.126550] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b985a88d-d593-4847-b93c-7d83fa86b5ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.159644] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56279d3-a0c6-446f-91f4-d6db18ecbbd6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.168120] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890b0030-de85-4593-8dc3-a63efec94df3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.183799] env[61728]: DEBUG nova.compute.provider_tree [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1148.336143] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465028, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.711105] env[61728]: DEBUG nova.scheduler.client.report [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 187 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1148.711413] env[61728]: DEBUG nova.compute.provider_tree [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 187 to 188 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1148.711620] env[61728]: DEBUG nova.compute.provider_tree [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1148.836354] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465028, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.111439} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.836651] env[61728]: INFO nova.virt.vmwareapi.ds_util [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] c92e595d-c686-4810-bb39-d98de4141121/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk. [ 1148.837474] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c1cba7-7596-498e-b41c-1dec918ffb7f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.862344] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Reconfiguring VM instance instance-00000077 to attach disk [datastore1] c92e595d-c686-4810-bb39-d98de4141121/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1148.862926] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72259a04-abcd-47fb-bdd3-71ba8cdc88f0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.881222] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1148.881222] env[61728]: value = "task-465029" [ 1148.881222] env[61728]: _type = "Task" [ 1148.881222] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.889627] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465029, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.216995] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.974s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.217546] env[61728]: DEBUG nova.compute.manager [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1149.222411] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.907s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.222411] env[61728]: DEBUG nova.objects.instance [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lazy-loading 'resources' on Instance uuid c1ed0e8b-5cf9-4ff0-931e-fadc3432a125 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1149.391248] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465029, 'name': ReconfigVM_Task, 'duration_secs': 0.291692} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.391572] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Reconfigured VM instance instance-00000077 to attach disk [datastore1] c92e595d-c686-4810-bb39-d98de4141121/8b767102-1435-4827-a43b-8e2e25ec780b-rescue.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1149.392379] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b9213f-0758-4db0-a15a-6bca3390f876 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.416136] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5b87375-bde7-4a9f-b084-a125dd9ed97a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.430481] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1149.430481] env[61728]: value = "task-465031" [ 1149.430481] env[61728]: _type = "Task" [ 1149.430481] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.440880] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465031, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.723821] env[61728]: DEBUG nova.compute.utils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1149.728560] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.729235] env[61728]: DEBUG nova.compute.manager [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1149.729322] env[61728]: DEBUG nova.network.neutron [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1149.772873] env[61728]: DEBUG nova.policy [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67d7c6a6191045cf85f790acebaf811e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7f61c87df8b485b9a6370032dd7af37', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1149.821123] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca7a4e5-5c29-4424-aa78-d76c6922ca66 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.830518] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a77b7dc8-4123-4f60-884c-b46bf24eecd4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.875448] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0094840c-2eac-45f7-a79c-428550fa68bb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.885692] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-058de653-28af-40b8-bdff-8373162fa509 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.900980] env[61728]: DEBUG nova.compute.provider_tree [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1149.940133] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465031, 'name': ReconfigVM_Task, 'duration_secs': 0.146967} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.940434] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1149.940703] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c3ad0b81-e960-4e79-9203-fe6aca860198 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.947644] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1149.947644] env[61728]: value = "task-465032" [ 1149.947644] env[61728]: _type = "Task" [ 1149.947644] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.955034] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465032, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.071347] env[61728]: DEBUG nova.network.neutron [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Successfully created port: 8787e957-2e8a-4cce-b359-84b4b077d388 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1150.230063] env[61728]: DEBUG nova.compute.manager [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1150.434868] env[61728]: DEBUG nova.scheduler.client.report [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 188 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1150.435172] env[61728]: DEBUG nova.compute.provider_tree [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 188 to 189 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1150.435366] env[61728]: DEBUG nova.compute.provider_tree [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1150.460231] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465032, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.941424] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.721s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.943862] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.215s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.944116] env[61728]: DEBUG nova.objects.instance [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lazy-loading 'pci_requests' on Instance uuid 532507fa-996d-45c7-bf70-f09c93be79ed {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1150.958606] env[61728]: DEBUG oslo_vmware.api [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465032, 'name': PowerOnVM_Task, 'duration_secs': 1.006111} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.959084] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1150.962111] env[61728]: DEBUG nova.compute.manager [None req-8d943e28-78b7-4f18-8f33-bb5eaba6863c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1150.962919] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af6a1cf6-724e-44a5-a1db-ac271e7123a9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.966172] env[61728]: INFO nova.scheduler.client.report [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Deleted allocations for instance c1ed0e8b-5cf9-4ff0-931e-fadc3432a125 [ 1151.241044] env[61728]: DEBUG nova.compute.manager [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1151.270787] env[61728]: DEBUG nova.virt.hardware [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1151.271111] env[61728]: DEBUG nova.virt.hardware [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1151.271288] env[61728]: DEBUG nova.virt.hardware [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1151.271623] env[61728]: DEBUG nova.virt.hardware [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1151.271854] env[61728]: DEBUG nova.virt.hardware [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1151.272052] env[61728]: DEBUG nova.virt.hardware [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1151.272331] env[61728]: DEBUG nova.virt.hardware [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1151.272545] env[61728]: DEBUG nova.virt.hardware [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1151.272770] env[61728]: DEBUG nova.virt.hardware [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1151.272986] env[61728]: DEBUG nova.virt.hardware [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1151.273199] env[61728]: DEBUG nova.virt.hardware [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1151.274370] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de3dbf0-422e-4c23-9d39-71c695285926 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.282744] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f202de-72a4-4faa-b749-d3ed14e68481 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.448321] env[61728]: DEBUG nova.objects.instance [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lazy-loading 'numa_topology' on Instance uuid 532507fa-996d-45c7-bf70-f09c93be79ed {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1151.477251] env[61728]: DEBUG oslo_concurrency.lockutils [None req-3548edf9-67d9-434e-a69e-e63b31a62e56 tempest-AttachVolumeNegativeTest-776470976 tempest-AttachVolumeNegativeTest-776470976-project-member] Lock "c1ed0e8b-5cf9-4ff0-931e-fadc3432a125" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.356s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.497340] env[61728]: DEBUG nova.compute.manager [req-44358b10-9323-4fc8-83ef-877a94a6c117 req-ff113eaa-b40a-4fde-99d2-dbdb9344256b service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Received event network-vif-plugged-8787e957-2e8a-4cce-b359-84b4b077d388 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1151.497604] env[61728]: DEBUG oslo_concurrency.lockutils [req-44358b10-9323-4fc8-83ef-877a94a6c117 req-ff113eaa-b40a-4fde-99d2-dbdb9344256b service nova] Acquiring lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.497823] env[61728]: DEBUG oslo_concurrency.lockutils [req-44358b10-9323-4fc8-83ef-877a94a6c117 req-ff113eaa-b40a-4fde-99d2-dbdb9344256b service nova] Lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.497997] env[61728]: DEBUG oslo_concurrency.lockutils [req-44358b10-9323-4fc8-83ef-877a94a6c117 req-ff113eaa-b40a-4fde-99d2-dbdb9344256b service nova] Lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.502578] env[61728]: DEBUG nova.compute.manager [req-44358b10-9323-4fc8-83ef-877a94a6c117 req-ff113eaa-b40a-4fde-99d2-dbdb9344256b service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] No waiting events found dispatching network-vif-plugged-8787e957-2e8a-4cce-b359-84b4b077d388 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1151.502863] env[61728]: WARNING nova.compute.manager [req-44358b10-9323-4fc8-83ef-877a94a6c117 req-ff113eaa-b40a-4fde-99d2-dbdb9344256b service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Received unexpected event network-vif-plugged-8787e957-2e8a-4cce-b359-84b4b077d388 for instance with vm_state building and task_state spawning. [ 1151.666928] env[61728]: DEBUG nova.network.neutron [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Successfully updated port: 8787e957-2e8a-4cce-b359-84b4b077d388 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1151.953186] env[61728]: INFO nova.compute.claims [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1152.045642] env[61728]: INFO nova.compute.manager [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Unrescuing [ 1152.046292] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "refresh_cache-c92e595d-c686-4810-bb39-d98de4141121" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.046292] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquired lock "refresh_cache-c92e595d-c686-4810-bb39-d98de4141121" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.046491] env[61728]: DEBUG nova.network.neutron [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1152.173736] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.173921] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.174199] env[61728]: DEBUG nova.network.neutron [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1152.703039] env[61728]: DEBUG nova.network.neutron [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1152.820790] env[61728]: DEBUG nova.network.neutron [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Updating instance_info_cache with network_info: [{"id": "7aee91de-db46-4071-8f9e-980bcfc29365", "address": "fa:16:3e:65:1f:da", "network": {"id": "4bb51519-042f-44ea-9325-a479be9790d3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1499510217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "909c86730f554a04ab05dcad3a53e9dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7aee91de-db", "ovs_interfaceid": "7aee91de-db46-4071-8f9e-980bcfc29365", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1152.874929] env[61728]: DEBUG nova.network.neutron [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance_info_cache with network_info: [{"id": "8787e957-2e8a-4cce-b359-84b4b077d388", "address": "fa:16:3e:d6:47:92", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8787e957-2e", "ovs_interfaceid": "8787e957-2e8a-4cce-b359-84b4b077d388", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.070072] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03ec9ff-51c6-4ae4-a40d-b4f459c9c029 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.078195] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70c79cb-3a59-443a-b45d-cc22691212d1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.111211] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63745b61-0733-4b16-9d65-440dfa417efd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.116823] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed2c9f5-b3b7-4afd-9be2-8df3f6ae205e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.130834] env[61728]: DEBUG nova.compute.provider_tree [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1153.323438] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Releasing lock "refresh_cache-c92e595d-c686-4810-bb39-d98de4141121" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.324423] env[61728]: DEBUG nova.objects.instance [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lazy-loading 'flavor' on Instance uuid c92e595d-c686-4810-bb39-d98de4141121 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1153.377951] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.379113] env[61728]: DEBUG nova.compute.manager [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Instance network_info: |[{"id": "8787e957-2e8a-4cce-b359-84b4b077d388", "address": "fa:16:3e:d6:47:92", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8787e957-2e", "ovs_interfaceid": "8787e957-2e8a-4cce-b359-84b4b077d388", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1153.379113] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:47:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b80dd748-3d7e-4a23-a38d-9e79a3881452', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8787e957-2e8a-4cce-b359-84b4b077d388', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1153.386633] env[61728]: DEBUG oslo.service.loopingcall [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1153.386888] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1153.387887] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7eab1f8d-dec3-424c-93a8-036a24c45620 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.409611] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1153.409611] env[61728]: value = "task-465036" [ 1153.409611] env[61728]: _type = "Task" [ 1153.409611] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.417402] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465036, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.524909] env[61728]: DEBUG nova.compute.manager [req-c2be973c-1d0e-4255-bb1b-9876e67a9f86 req-208eaaca-8193-46ca-890d-fad1ea48d3b0 service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Received event network-changed-8787e957-2e8a-4cce-b359-84b4b077d388 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1153.525024] env[61728]: DEBUG nova.compute.manager [req-c2be973c-1d0e-4255-bb1b-9876e67a9f86 req-208eaaca-8193-46ca-890d-fad1ea48d3b0 service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Refreshing instance network info cache due to event network-changed-8787e957-2e8a-4cce-b359-84b4b077d388. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1153.525322] env[61728]: DEBUG oslo_concurrency.lockutils [req-c2be973c-1d0e-4255-bb1b-9876e67a9f86 req-208eaaca-8193-46ca-890d-fad1ea48d3b0 service nova] Acquiring lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.525535] env[61728]: DEBUG oslo_concurrency.lockutils [req-c2be973c-1d0e-4255-bb1b-9876e67a9f86 req-208eaaca-8193-46ca-890d-fad1ea48d3b0 service nova] Acquired lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.525697] env[61728]: DEBUG nova.network.neutron [req-c2be973c-1d0e-4255-bb1b-9876e67a9f86 req-208eaaca-8193-46ca-890d-fad1ea48d3b0 service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Refreshing network info cache for port 8787e957-2e8a-4cce-b359-84b4b077d388 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1153.634636] env[61728]: DEBUG nova.scheduler.client.report [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1153.831917] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392a86ba-6e15-4eef-9c07-552862407e14 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.855619] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1153.855995] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1dd04080-dfea-473f-86e9-06a62a5e8c05 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.862449] env[61728]: DEBUG oslo_vmware.api [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1153.862449] env[61728]: value = "task-465037" [ 1153.862449] env[61728]: _type = "Task" [ 1153.862449] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.870833] env[61728]: DEBUG oslo_vmware.api [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.922146] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465036, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.140556] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.197s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.173962] env[61728]: INFO nova.network.neutron [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updating port d04bc3a5-d378-4e1e-a145-ec523bcc442a with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1154.374480] env[61728]: DEBUG oslo_vmware.api [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465037, 'name': PowerOffVM_Task, 'duration_secs': 0.245309} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.374846] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1154.380965] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Reconfiguring VM instance instance-00000077 to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1154.381321] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e34e2bd-01b1-499f-a059-c79873c53537 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.404086] env[61728]: DEBUG oslo_vmware.api [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1154.404086] env[61728]: value = "task-465038" [ 1154.404086] env[61728]: _type = "Task" [ 1154.404086] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.413508] env[61728]: DEBUG oslo_vmware.api [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465038, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.423024] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465036, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.674190] env[61728]: DEBUG nova.network.neutron [req-c2be973c-1d0e-4255-bb1b-9876e67a9f86 req-208eaaca-8193-46ca-890d-fad1ea48d3b0 service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updated VIF entry in instance network info cache for port 8787e957-2e8a-4cce-b359-84b4b077d388. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1154.674576] env[61728]: DEBUG nova.network.neutron [req-c2be973c-1d0e-4255-bb1b-9876e67a9f86 req-208eaaca-8193-46ca-890d-fad1ea48d3b0 service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance_info_cache with network_info: [{"id": "8787e957-2e8a-4cce-b359-84b4b077d388", "address": "fa:16:3e:d6:47:92", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8787e957-2e", "ovs_interfaceid": "8787e957-2e8a-4cce-b359-84b4b077d388", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.914776] env[61728]: DEBUG oslo_vmware.api [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465038, 'name': ReconfigVM_Task, 'duration_secs': 0.210676} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.920796] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Reconfigured VM instance instance-00000077 to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1154.920796] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1154.920796] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6aaabd4-2a7f-44fd-ad63-b9acce9c70c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.927683] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465036, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.927683] env[61728]: DEBUG oslo_vmware.api [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1154.927683] env[61728]: value = "task-465040" [ 1154.927683] env[61728]: _type = "Task" [ 1154.927683] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.934173] env[61728]: DEBUG oslo_vmware.api [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465040, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.177246] env[61728]: DEBUG oslo_concurrency.lockutils [req-c2be973c-1d0e-4255-bb1b-9876e67a9f86 req-208eaaca-8193-46ca-890d-fad1ea48d3b0 service nova] Releasing lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1155.426995] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465036, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.435752] env[61728]: DEBUG oslo_vmware.api [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465040, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.630877] env[61728]: DEBUG nova.compute.manager [req-b786ccb9-8bb0-428b-810f-3761c921ce0d req-f95dfeea-7e6f-4bc1-af77-91e9f3f0e870 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Received event network-vif-plugged-d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1155.631136] env[61728]: DEBUG oslo_concurrency.lockutils [req-b786ccb9-8bb0-428b-810f-3761c921ce0d req-f95dfeea-7e6f-4bc1-af77-91e9f3f0e870 service nova] Acquiring lock "532507fa-996d-45c7-bf70-f09c93be79ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.631362] env[61728]: DEBUG oslo_concurrency.lockutils [req-b786ccb9-8bb0-428b-810f-3761c921ce0d req-f95dfeea-7e6f-4bc1-af77-91e9f3f0e870 service nova] Lock "532507fa-996d-45c7-bf70-f09c93be79ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.632580] env[61728]: DEBUG oslo_concurrency.lockutils [req-b786ccb9-8bb0-428b-810f-3761c921ce0d req-f95dfeea-7e6f-4bc1-af77-91e9f3f0e870 service nova] Lock "532507fa-996d-45c7-bf70-f09c93be79ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.632630] env[61728]: DEBUG nova.compute.manager [req-b786ccb9-8bb0-428b-810f-3761c921ce0d req-f95dfeea-7e6f-4bc1-af77-91e9f3f0e870 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] No waiting events found dispatching network-vif-plugged-d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1155.632828] env[61728]: WARNING nova.compute.manager [req-b786ccb9-8bb0-428b-810f-3761c921ce0d req-f95dfeea-7e6f-4bc1-af77-91e9f3f0e870 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Received unexpected event network-vif-plugged-d04bc3a5-d378-4e1e-a145-ec523bcc442a for instance with vm_state shelved_offloaded and task_state spawning. [ 1155.669628] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.669628] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.669743] env[61728]: DEBUG nova.network.neutron [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1155.741384] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Acquiring lock "76429588-fb4b-4215-b78c-75ce366989ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.741483] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Lock "76429588-fb4b-4215-b78c-75ce366989ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.925682] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465036, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.936452] env[61728]: DEBUG oslo_vmware.api [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465040, 'name': PowerOnVM_Task, 'duration_secs': 0.665218} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.936750] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1155.936988] env[61728]: DEBUG nova.compute.manager [None req-c059338d-7bec-404e-97c1-390ec3ad6782 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1155.937860] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07a3f23-9064-4b4c-8f39-b9f796674cfc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.243608] env[61728]: DEBUG nova.compute.manager [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1156.387334] env[61728]: DEBUG nova.network.neutron [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updating instance_info_cache with network_info: [{"id": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "address": "fa:16:3e:f1:65:32", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04bc3a5-d3", "ovs_interfaceid": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.425489] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465036, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.766057] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.766345] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.768026] env[61728]: INFO nova.compute.claims [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1156.891592] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.921378] env[61728]: DEBUG nova.virt.hardware [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='5bc1c53946ff77a6a5a72e1e4bc13f36',container_format='bare',created_at=2024-08-29T12:32:54Z,direct_url=,disk_format='vmdk',id=df2f6391-a88d-4bb9-ada0-dd01a6acec62,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-149515067-shelved',owner='438de94d086e47b99cedc6e4a33471bf',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-08-29T12:33:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1156.921638] env[61728]: DEBUG nova.virt.hardware [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1156.921887] env[61728]: DEBUG nova.virt.hardware [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1156.922151] env[61728]: DEBUG nova.virt.hardware [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1156.922622] env[61728]: DEBUG nova.virt.hardware [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1156.922622] env[61728]: DEBUG nova.virt.hardware [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1156.922747] env[61728]: DEBUG nova.virt.hardware [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1156.922889] env[61728]: DEBUG nova.virt.hardware [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1156.923101] env[61728]: DEBUG nova.virt.hardware [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1156.923433] env[61728]: DEBUG nova.virt.hardware [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1156.923433] env[61728]: DEBUG nova.virt.hardware [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1156.924284] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e993b997-5ecc-44f0-9f7e-24dd994247a3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.934575] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465036, 'name': CreateVM_Task, 'duration_secs': 3.097866} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.934830] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1156.936072] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a6f67b0-2359-49cd-9b42-93debc39d1c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.940049] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.940234] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.940552] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1156.940796] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e53a5fed-9c76-4db3-b692-ec151d7d50bf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.945964] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1156.945964] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521434d2-625f-6fda-f578-90d420d6f263" [ 1156.945964] env[61728]: _type = "Task" [ 1156.945964] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.954250] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:65:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f856fca-9fb5-41ea-a057-ac4193bd323d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd04bc3a5-d378-4e1e-a145-ec523bcc442a', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1156.962094] env[61728]: DEBUG oslo.service.loopingcall [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1156.965601] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1156.965827] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a7a683fd-5721-4f48-8120-b6722e73b976 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.985300] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]521434d2-625f-6fda-f578-90d420d6f263, 'name': SearchDatastore_Task, 'duration_secs': 0.010333} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.986554] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.986788] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1156.987118] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.987292] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.987481] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1156.987696] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1156.987696] env[61728]: value = "task-465041" [ 1156.987696] env[61728]: _type = "Task" [ 1156.987696] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.987873] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4689dc18-1f13-4fc8-840a-2c83b23953ec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.997051] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465041, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.998149] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1156.998335] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1156.999062] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05fd7f41-1e56-42a8-8b01-a181587ffa05 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.003832] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1157.003832] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528afda0-aa0b-c297-b3a8-5c4ad9c6d0c7" [ 1157.003832] env[61728]: _type = "Task" [ 1157.003832] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.011311] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528afda0-aa0b-c297-b3a8-5c4ad9c6d0c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.330603] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "c92e595d-c686-4810-bb39-d98de4141121" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.330765] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "c92e595d-c686-4810-bb39-d98de4141121" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.331048] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "c92e595d-c686-4810-bb39-d98de4141121-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.332056] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "c92e595d-c686-4810-bb39-d98de4141121-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.332056] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "c92e595d-c686-4810-bb39-d98de4141121-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.334112] env[61728]: INFO nova.compute.manager [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Terminating instance [ 1157.336230] env[61728]: DEBUG nova.compute.manager [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1157.336478] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1157.337428] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0a506e-d4f2-4ed7-82b8-e6a2d85d328f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.346790] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1157.347039] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-654e8502-1872-4167-975f-e05b16ae35cf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.354382] env[61728]: DEBUG oslo_vmware.api [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1157.354382] env[61728]: value = "task-465043" [ 1157.354382] env[61728]: _type = "Task" [ 1157.354382] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.362549] env[61728]: DEBUG oslo_vmware.api [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465043, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.499626] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465041, 'name': CreateVM_Task, 'duration_secs': 0.32451} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.499816] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1157.500449] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/df2f6391-a88d-4bb9-ada0-dd01a6acec62" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.500632] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired lock "[datastore1] devstack-image-cache_base/df2f6391-a88d-4bb9-ada0-dd01a6acec62" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.501040] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/df2f6391-a88d-4bb9-ada0-dd01a6acec62" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1157.501318] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d8c32fd-5ee0-48ac-bc3e-37c13935efd2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.508487] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1157.508487] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520daeb1-b9a6-aa03-ced6-8a8c8f39743f" [ 1157.508487] env[61728]: _type = "Task" [ 1157.508487] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.515322] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528afda0-aa0b-c297-b3a8-5c4ad9c6d0c7, 'name': SearchDatastore_Task, 'duration_secs': 0.008566} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.516468] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e6aa111-d22d-4caf-8e5b-e8b2c9dce63e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.521638] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520daeb1-b9a6-aa03-ced6-8a8c8f39743f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.525707] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1157.525707] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bd6378-976a-3876-4504-ae20f07e925c" [ 1157.525707] env[61728]: _type = "Task" [ 1157.525707] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.536793] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bd6378-976a-3876-4504-ae20f07e925c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.654524] env[61728]: DEBUG nova.compute.manager [req-b2e97276-6b5a-4694-8c37-0a44f3e485a8 req-2e985e10-d088-4cf7-8b20-6c3cc18716bd service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Received event network-changed-d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1157.654733] env[61728]: DEBUG nova.compute.manager [req-b2e97276-6b5a-4694-8c37-0a44f3e485a8 req-2e985e10-d088-4cf7-8b20-6c3cc18716bd service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Refreshing instance network info cache due to event network-changed-d04bc3a5-d378-4e1e-a145-ec523bcc442a. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1157.654954] env[61728]: DEBUG oslo_concurrency.lockutils [req-b2e97276-6b5a-4694-8c37-0a44f3e485a8 req-2e985e10-d088-4cf7-8b20-6c3cc18716bd service nova] Acquiring lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.655272] env[61728]: DEBUG oslo_concurrency.lockutils [req-b2e97276-6b5a-4694-8c37-0a44f3e485a8 req-2e985e10-d088-4cf7-8b20-6c3cc18716bd service nova] Acquired lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.655466] env[61728]: DEBUG nova.network.neutron [req-b2e97276-6b5a-4694-8c37-0a44f3e485a8 req-2e985e10-d088-4cf7-8b20-6c3cc18716bd service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Refreshing network info cache for port d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1157.867244] env[61728]: DEBUG oslo_vmware.api [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465043, 'name': PowerOffVM_Task, 'duration_secs': 0.188038} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.867537] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1157.867712] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1157.867980] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cdc68e98-c4f7-4105-94bd-27abb167e19c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.883412] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8512e4e3-3890-4124-a972-adf805b713cf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.890737] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f005cd32-fd59-4e42-a71b-db6cd0413c0f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.922251] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c5369e-1001-4266-b198-953afe167bd3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.931327] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb0a9c5d-7de4-42b7-8a2a-752428ad5394 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.937147] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1157.937527] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1157.937801] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Deleting the datastore file [datastore1] c92e595d-c686-4810-bb39-d98de4141121 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1157.938620] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9bdef844-645b-4c7a-825b-77ae5c405c09 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.950015] env[61728]: DEBUG nova.compute.provider_tree [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1157.953334] env[61728]: DEBUG oslo_vmware.api [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1157.953334] env[61728]: value = "task-465046" [ 1157.953334] env[61728]: _type = "Task" [ 1157.953334] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.962502] env[61728]: DEBUG oslo_vmware.api [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465046, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.020104] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lock "[datastore1] devstack-image-cache_base/df2f6391-a88d-4bb9-ada0-dd01a6acec62" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.020409] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Processing image df2f6391-a88d-4bb9-ada0-dd01a6acec62 {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1158.020936] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/df2f6391-a88d-4bb9-ada0-dd01a6acec62/df2f6391-a88d-4bb9-ada0-dd01a6acec62.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.021044] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired lock "[datastore1] devstack-image-cache_base/df2f6391-a88d-4bb9-ada0-dd01a6acec62/df2f6391-a88d-4bb9-ada0-dd01a6acec62.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.021267] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1158.021621] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2e24ef8-03eb-46c3-9c54-9bdc038dcdd0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.031144] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1158.031355] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1158.032484] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8baf06ed-6ae6-488d-be16-3c45c83e9724 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.039942] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52bd6378-976a-3876-4504-ae20f07e925c, 'name': SearchDatastore_Task, 'duration_secs': 0.01876} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.040544] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.040837] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] fb1195b2-1ab5-47d7-be86-4b61e47b4e29/fb1195b2-1ab5-47d7-be86-4b61e47b4e29.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1158.041115] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba51f9a3-7abe-4f53-a14c-d10722968a34 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.044225] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1158.044225] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5284a053-26f4-3f52-8f98-193bb95d337e" [ 1158.044225] env[61728]: _type = "Task" [ 1158.044225] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.050200] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1158.050200] env[61728]: value = "task-465047" [ 1158.050200] env[61728]: _type = "Task" [ 1158.050200] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.053398] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5284a053-26f4-3f52-8f98-193bb95d337e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.062284] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465047, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.386538] env[61728]: DEBUG nova.network.neutron [req-b2e97276-6b5a-4694-8c37-0a44f3e485a8 req-2e985e10-d088-4cf7-8b20-6c3cc18716bd service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updated VIF entry in instance network info cache for port d04bc3a5-d378-4e1e-a145-ec523bcc442a. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1158.386963] env[61728]: DEBUG nova.network.neutron [req-b2e97276-6b5a-4694-8c37-0a44f3e485a8 req-2e985e10-d088-4cf7-8b20-6c3cc18716bd service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updating instance_info_cache with network_info: [{"id": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "address": "fa:16:3e:f1:65:32", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04bc3a5-d3", "ovs_interfaceid": "d04bc3a5-d378-4e1e-a145-ec523bcc442a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.453561] env[61728]: DEBUG nova.scheduler.client.report [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1158.468488] env[61728]: DEBUG oslo_vmware.api [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465046, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15664} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.468776] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1158.468962] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1158.469162] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1158.469378] env[61728]: INFO nova.compute.manager [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: c92e595d-c686-4810-bb39-d98de4141121] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1158.469731] env[61728]: DEBUG oslo.service.loopingcall [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1158.470541] env[61728]: DEBUG nova.compute.manager [-] [instance: c92e595d-c686-4810-bb39-d98de4141121] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1158.470541] env[61728]: DEBUG nova.network.neutron [-] [instance: c92e595d-c686-4810-bb39-d98de4141121] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1158.557411] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Preparing fetch location {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1158.557671] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Fetch image to [datastore1] OSTACK_IMG_cf5a39d3-2fef-49d4-ac9f-06efde512f0b/OSTACK_IMG_cf5a39d3-2fef-49d4-ac9f-06efde512f0b.vmdk {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1158.557854] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Downloading stream optimized image df2f6391-a88d-4bb9-ada0-dd01a6acec62 to [datastore1] OSTACK_IMG_cf5a39d3-2fef-49d4-ac9f-06efde512f0b/OSTACK_IMG_cf5a39d3-2fef-49d4-ac9f-06efde512f0b.vmdk on the data store datastore1 as vApp {{(pid=61728) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1158.558041] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Downloading image file data df2f6391-a88d-4bb9-ada0-dd01a6acec62 to the ESX as VM named 'OSTACK_IMG_cf5a39d3-2fef-49d4-ac9f-06efde512f0b' {{(pid=61728) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1158.564916] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465047, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45108} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.565179] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] fb1195b2-1ab5-47d7-be86-4b61e47b4e29/fb1195b2-1ab5-47d7-be86-4b61e47b4e29.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1158.565389] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1158.565627] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-477a2692-7777-460b-9c28-906b62fbd9aa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.572143] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1158.572143] env[61728]: value = "task-465048" [ 1158.572143] env[61728]: _type = "Task" [ 1158.572143] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.580302] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465048, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.635441] env[61728]: DEBUG oslo_vmware.rw_handles [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1158.635441] env[61728]: value = "resgroup-9" [ 1158.635441] env[61728]: _type = "ResourcePool" [ 1158.635441] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1158.635767] env[61728]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-0154c689-554d-4ea0-bcab-a29603f6eb55 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.656944] env[61728]: DEBUG oslo_vmware.rw_handles [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lease: (returnval){ [ 1158.656944] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523f4632-c71f-564b-38e7-a45f86fb9eba" [ 1158.656944] env[61728]: _type = "HttpNfcLease" [ 1158.656944] env[61728]: } obtained for vApp import into resource pool (val){ [ 1158.656944] env[61728]: value = "resgroup-9" [ 1158.656944] env[61728]: _type = "ResourcePool" [ 1158.656944] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1158.657355] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the lease: (returnval){ [ 1158.657355] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523f4632-c71f-564b-38e7-a45f86fb9eba" [ 1158.657355] env[61728]: _type = "HttpNfcLease" [ 1158.657355] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1158.669747] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1158.669747] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523f4632-c71f-564b-38e7-a45f86fb9eba" [ 1158.669747] env[61728]: _type = "HttpNfcLease" [ 1158.669747] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1158.890735] env[61728]: DEBUG oslo_concurrency.lockutils [req-b2e97276-6b5a-4694-8c37-0a44f3e485a8 req-2e985e10-d088-4cf7-8b20-6c3cc18716bd service nova] Releasing lock "refresh_cache-532507fa-996d-45c7-bf70-f09c93be79ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.963473] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.197s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.964216] env[61728]: DEBUG nova.compute.manager [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1159.082554] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465048, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060125} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.082853] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1159.083738] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33adec6e-8796-4fe6-8ba5-875d5f560bde {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.119369] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Reconfiguring VM instance instance-00000078 to attach disk [datastore1] fb1195b2-1ab5-47d7-be86-4b61e47b4e29/fb1195b2-1ab5-47d7-be86-4b61e47b4e29.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1159.119677] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1697f566-237d-469f-a2a1-0b70ad3dc69f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.145089] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1159.145089] env[61728]: value = "task-465050" [ 1159.145089] env[61728]: _type = "Task" [ 1159.145089] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.163944] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465050, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.171521] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1159.171521] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523f4632-c71f-564b-38e7-a45f86fb9eba" [ 1159.171521] env[61728]: _type = "HttpNfcLease" [ 1159.171521] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1159.205132] env[61728]: DEBUG nova.network.neutron [-] [instance: c92e595d-c686-4810-bb39-d98de4141121] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.469797] env[61728]: DEBUG nova.compute.utils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1159.471488] env[61728]: DEBUG nova.compute.manager [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1159.471659] env[61728]: DEBUG nova.network.neutron [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1159.522909] env[61728]: DEBUG nova.policy [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '286aae23614643f2a224af9bda89d0dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd34b224bdc2149288acb934783eb9954', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1159.655630] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465050, 'name': ReconfigVM_Task, 'duration_secs': 0.295633} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.656058] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Reconfigured VM instance instance-00000078 to attach disk [datastore1] fb1195b2-1ab5-47d7-be86-4b61e47b4e29/fb1195b2-1ab5-47d7-be86-4b61e47b4e29.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1159.656646] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-beb16bfc-f433-4f6a-b37c-246d0c8a361b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.663492] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1159.663492] env[61728]: value = "task-465052" [ 1159.663492] env[61728]: _type = "Task" [ 1159.663492] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.673133] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1159.673133] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523f4632-c71f-564b-38e7-a45f86fb9eba" [ 1159.673133] env[61728]: _type = "HttpNfcLease" [ 1159.673133] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1159.676319] env[61728]: DEBUG oslo_vmware.rw_handles [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1159.676319] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]523f4632-c71f-564b-38e7-a45f86fb9eba" [ 1159.676319] env[61728]: _type = "HttpNfcLease" [ 1159.676319] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1159.676614] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465052, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.677384] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815153d6-d923-41ed-af00-e81a972814c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.681637] env[61728]: DEBUG nova.compute.manager [req-691d2bd5-bc58-4d4a-9179-ed7f41cf8f2d req-f2326921-f03e-4f4a-a231-9cc21f2d5fbb service nova] [instance: c92e595d-c686-4810-bb39-d98de4141121] Received event network-vif-deleted-7aee91de-db46-4071-8f9e-980bcfc29365 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1159.688013] env[61728]: DEBUG oslo_vmware.rw_handles [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5206819c-ab79-f3dd-8d4f-e023254d2997/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1159.688213] env[61728]: DEBUG oslo_vmware.rw_handles [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5206819c-ab79-f3dd-8d4f-e023254d2997/disk-0.vmdk. {{(pid=61728) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1159.745425] env[61728]: INFO nova.compute.manager [-] [instance: c92e595d-c686-4810-bb39-d98de4141121] Took 1.27 seconds to deallocate network for instance. [ 1159.756900] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4c943a6a-18db-4480-98a6-5047baa57649 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.773045] env[61728]: DEBUG nova.network.neutron [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Successfully created port: 704e9c0a-f4ff-48f7-a48f-ff4073e5929e {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1159.975197] env[61728]: DEBUG nova.compute.manager [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1160.176995] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465052, 'name': Rename_Task, 'duration_secs': 0.151748} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.178986] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1160.180563] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3fa5f85a-1b4c-4098-9143-e3a7132a827d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.188262] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1160.188262] env[61728]: value = "task-465053" [ 1160.188262] env[61728]: _type = "Task" [ 1160.188262] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.198591] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465053, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.255640] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.255924] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.257017] env[61728]: DEBUG nova.objects.instance [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lazy-loading 'resources' on Instance uuid c92e595d-c686-4810-bb39-d98de4141121 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1160.703136] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465053, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.884172] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d7170dd-0405-484f-825d-dbdf3a193584 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.894661] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50769274-9a69-4dd1-a5b3-4e28f304e771 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.927868] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24cde98-1d59-4f21-81cd-b17532e333c5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.937707] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d48fb20-2b43-473a-944d-7d8e996b3fa0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.952877] env[61728]: DEBUG nova.compute.provider_tree [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1160.985578] env[61728]: DEBUG nova.compute.manager [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1161.016338] env[61728]: DEBUG nova.virt.hardware [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1161.016667] env[61728]: DEBUG nova.virt.hardware [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1161.016936] env[61728]: DEBUG nova.virt.hardware [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1161.017135] env[61728]: DEBUG nova.virt.hardware [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1161.017235] env[61728]: DEBUG nova.virt.hardware [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1161.017382] env[61728]: DEBUG nova.virt.hardware [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1161.017616] env[61728]: DEBUG nova.virt.hardware [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1161.017823] env[61728]: DEBUG nova.virt.hardware [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1161.018036] env[61728]: DEBUG nova.virt.hardware [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1161.018222] env[61728]: DEBUG nova.virt.hardware [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1161.018427] env[61728]: DEBUG nova.virt.hardware [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1161.019441] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd702f0-27af-4621-afb5-ba8acfd55de1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.030950] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da145f8f-e38c-4f74-b84b-d83b28964c9b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.128103] env[61728]: DEBUG oslo_vmware.rw_handles [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Completed reading data from the image iterator. {{(pid=61728) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1161.128331] env[61728]: DEBUG oslo_vmware.rw_handles [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5206819c-ab79-f3dd-8d4f-e023254d2997/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1161.129508] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e928d742-47b8-497e-ae7b-9378e61bf528 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.138460] env[61728]: DEBUG oslo_vmware.rw_handles [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5206819c-ab79-f3dd-8d4f-e023254d2997/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1161.138653] env[61728]: DEBUG oslo_vmware.rw_handles [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5206819c-ab79-f3dd-8d4f-e023254d2997/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1161.138919] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-d3ee08f1-c4ed-432e-83c0-16d9367b3056 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.202509] env[61728]: DEBUG oslo_vmware.api [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465053, 'name': PowerOnVM_Task, 'duration_secs': 0.517076} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.203431] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1161.203789] env[61728]: INFO nova.compute.manager [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Took 9.96 seconds to spawn the instance on the hypervisor. [ 1161.204165] env[61728]: DEBUG nova.compute.manager [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1161.205356] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8668c3fa-4767-4105-8084-dccdbf08fd73 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.365406] env[61728]: DEBUG oslo_vmware.rw_handles [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5206819c-ab79-f3dd-8d4f-e023254d2997/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1161.365406] env[61728]: INFO nova.virt.vmwareapi.images [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Downloaded image file data df2f6391-a88d-4bb9-ada0-dd01a6acec62 [ 1161.365406] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932172b8-9c65-460a-9d5a-99560a62c187 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.381103] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e55d7e43-be01-442c-9174-5c865f403a0f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.415141] env[61728]: INFO nova.virt.vmwareapi.images [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] The imported VM was unregistered [ 1161.417636] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Caching image {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1161.417888] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Creating directory with path [datastore1] devstack-image-cache_base/df2f6391-a88d-4bb9-ada0-dd01a6acec62 {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1161.418189] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-edf55730-a196-4e8e-b854-2db7645c8cd4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.430937] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Created directory with path [datastore1] devstack-image-cache_base/df2f6391-a88d-4bb9-ada0-dd01a6acec62 {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1161.430937] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_cf5a39d3-2fef-49d4-ac9f-06efde512f0b/OSTACK_IMG_cf5a39d3-2fef-49d4-ac9f-06efde512f0b.vmdk to [datastore1] devstack-image-cache_base/df2f6391-a88d-4bb9-ada0-dd01a6acec62/df2f6391-a88d-4bb9-ada0-dd01a6acec62.vmdk. {{(pid=61728) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1161.430937] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-a3731a13-3027-4b8b-8942-95815f845eb5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.438754] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1161.438754] env[61728]: value = "task-465056" [ 1161.438754] env[61728]: _type = "Task" [ 1161.438754] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.447465] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465056, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.457567] env[61728]: DEBUG nova.scheduler.client.report [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1161.630720] env[61728]: DEBUG nova.network.neutron [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Successfully updated port: 704e9c0a-f4ff-48f7-a48f-ff4073e5929e {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1161.717318] env[61728]: DEBUG nova.compute.manager [req-4f41d7b6-bb59-4ebd-9ae6-e9d40b39764b req-02a858fa-fb4d-4d07-8003-390464141d9d service nova] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Received event network-vif-plugged-704e9c0a-f4ff-48f7-a48f-ff4073e5929e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1161.717558] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f41d7b6-bb59-4ebd-9ae6-e9d40b39764b req-02a858fa-fb4d-4d07-8003-390464141d9d service nova] Acquiring lock "76429588-fb4b-4215-b78c-75ce366989ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.717780] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f41d7b6-bb59-4ebd-9ae6-e9d40b39764b req-02a858fa-fb4d-4d07-8003-390464141d9d service nova] Lock "76429588-fb4b-4215-b78c-75ce366989ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.717958] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f41d7b6-bb59-4ebd-9ae6-e9d40b39764b req-02a858fa-fb4d-4d07-8003-390464141d9d service nova] Lock "76429588-fb4b-4215-b78c-75ce366989ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.718215] env[61728]: DEBUG nova.compute.manager [req-4f41d7b6-bb59-4ebd-9ae6-e9d40b39764b req-02a858fa-fb4d-4d07-8003-390464141d9d service nova] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] No waiting events found dispatching network-vif-plugged-704e9c0a-f4ff-48f7-a48f-ff4073e5929e {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1161.718398] env[61728]: WARNING nova.compute.manager [req-4f41d7b6-bb59-4ebd-9ae6-e9d40b39764b req-02a858fa-fb4d-4d07-8003-390464141d9d service nova] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Received unexpected event network-vif-plugged-704e9c0a-f4ff-48f7-a48f-ff4073e5929e for instance with vm_state building and task_state spawning. [ 1161.718649] env[61728]: DEBUG nova.compute.manager [req-4f41d7b6-bb59-4ebd-9ae6-e9d40b39764b req-02a858fa-fb4d-4d07-8003-390464141d9d service nova] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Received event network-changed-704e9c0a-f4ff-48f7-a48f-ff4073e5929e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1161.718861] env[61728]: DEBUG nova.compute.manager [req-4f41d7b6-bb59-4ebd-9ae6-e9d40b39764b req-02a858fa-fb4d-4d07-8003-390464141d9d service nova] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Refreshing instance network info cache due to event network-changed-704e9c0a-f4ff-48f7-a48f-ff4073e5929e. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1161.719432] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f41d7b6-bb59-4ebd-9ae6-e9d40b39764b req-02a858fa-fb4d-4d07-8003-390464141d9d service nova] Acquiring lock "refresh_cache-76429588-fb4b-4215-b78c-75ce366989ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1161.719432] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f41d7b6-bb59-4ebd-9ae6-e9d40b39764b req-02a858fa-fb4d-4d07-8003-390464141d9d service nova] Acquired lock "refresh_cache-76429588-fb4b-4215-b78c-75ce366989ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.719432] env[61728]: DEBUG nova.network.neutron [req-4f41d7b6-bb59-4ebd-9ae6-e9d40b39764b req-02a858fa-fb4d-4d07-8003-390464141d9d service nova] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Refreshing network info cache for port 704e9c0a-f4ff-48f7-a48f-ff4073e5929e {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1161.733059] env[61728]: INFO nova.compute.manager [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Took 16.85 seconds to build instance. [ 1161.951832] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465056, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.963324] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.707s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.990368] env[61728]: INFO nova.scheduler.client.report [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Deleted allocations for instance c92e595d-c686-4810-bb39-d98de4141121 [ 1161.993759] env[61728]: DEBUG nova.compute.manager [req-4f5b5c8d-df2a-4e64-9824-6a24bf632f90 req-6d5025a0-ff68-48da-8261-39109646eb24 service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Received event network-changed-8787e957-2e8a-4cce-b359-84b4b077d388 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1161.994134] env[61728]: DEBUG nova.compute.manager [req-4f5b5c8d-df2a-4e64-9824-6a24bf632f90 req-6d5025a0-ff68-48da-8261-39109646eb24 service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Refreshing instance network info cache due to event network-changed-8787e957-2e8a-4cce-b359-84b4b077d388. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1161.994794] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f5b5c8d-df2a-4e64-9824-6a24bf632f90 req-6d5025a0-ff68-48da-8261-39109646eb24 service nova] Acquiring lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1161.994794] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f5b5c8d-df2a-4e64-9824-6a24bf632f90 req-6d5025a0-ff68-48da-8261-39109646eb24 service nova] Acquired lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.995081] env[61728]: DEBUG nova.network.neutron [req-4f5b5c8d-df2a-4e64-9824-6a24bf632f90 req-6d5025a0-ff68-48da-8261-39109646eb24 service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Refreshing network info cache for port 8787e957-2e8a-4cce-b359-84b4b077d388 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1162.133588] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Acquiring lock "refresh_cache-76429588-fb4b-4215-b78c-75ce366989ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1162.235386] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5b55ba33-f0d6-4487-b490-93f77480a9ff tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.359s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.269639] env[61728]: DEBUG nova.network.neutron [req-4f41d7b6-bb59-4ebd-9ae6-e9d40b39764b req-02a858fa-fb4d-4d07-8003-390464141d9d service nova] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1162.424328] env[61728]: DEBUG nova.network.neutron [req-4f41d7b6-bb59-4ebd-9ae6-e9d40b39764b req-02a858fa-fb4d-4d07-8003-390464141d9d service nova] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.449754] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465056, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.506015] env[61728]: DEBUG oslo_concurrency.lockutils [None req-4866c09a-d9be-4cc6-9518-343da3454642 tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "c92e595d-c686-4810-bb39-d98de4141121" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.175s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.790349] env[61728]: DEBUG nova.network.neutron [req-4f5b5c8d-df2a-4e64-9824-6a24bf632f90 req-6d5025a0-ff68-48da-8261-39109646eb24 service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updated VIF entry in instance network info cache for port 8787e957-2e8a-4cce-b359-84b4b077d388. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1162.790756] env[61728]: DEBUG nova.network.neutron [req-4f5b5c8d-df2a-4e64-9824-6a24bf632f90 req-6d5025a0-ff68-48da-8261-39109646eb24 service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance_info_cache with network_info: [{"id": "8787e957-2e8a-4cce-b359-84b4b077d388", "address": "fa:16:3e:d6:47:92", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8787e957-2e", "ovs_interfaceid": "8787e957-2e8a-4cce-b359-84b4b077d388", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.927700] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f41d7b6-bb59-4ebd-9ae6-e9d40b39764b req-02a858fa-fb4d-4d07-8003-390464141d9d service nova] Releasing lock "refresh_cache-76429588-fb4b-4215-b78c-75ce366989ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1162.928202] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Acquired lock "refresh_cache-76429588-fb4b-4215-b78c-75ce366989ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.928425] env[61728]: DEBUG nova.network.neutron [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1162.952839] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465056, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.956526] env[61728]: DEBUG oslo_concurrency.lockutils [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "79782d2b-1b90-4f3f-a1ce-394e88a00546" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.956719] env[61728]: DEBUG oslo_concurrency.lockutils [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "79782d2b-1b90-4f3f-a1ce-394e88a00546" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.956996] env[61728]: DEBUG oslo_concurrency.lockutils [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "79782d2b-1b90-4f3f-a1ce-394e88a00546-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.957222] env[61728]: DEBUG oslo_concurrency.lockutils [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "79782d2b-1b90-4f3f-a1ce-394e88a00546-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.957403] env[61728]: DEBUG oslo_concurrency.lockutils [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "79782d2b-1b90-4f3f-a1ce-394e88a00546-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.960412] env[61728]: INFO nova.compute.manager [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Terminating instance [ 1162.964193] env[61728]: DEBUG nova.compute.manager [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1162.964193] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1162.964916] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a4c86d-0f3b-4f2f-8549-08aeae416d0f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.975433] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1162.975811] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bac8496f-f3da-4c3c-ab0b-889eb23a192b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.983814] env[61728]: DEBUG oslo_vmware.api [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1162.983814] env[61728]: value = "task-465057" [ 1162.983814] env[61728]: _type = "Task" [ 1162.983814] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.994025] env[61728]: DEBUG oslo_vmware.api [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465057, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.294459] env[61728]: DEBUG oslo_concurrency.lockutils [req-4f5b5c8d-df2a-4e64-9824-6a24bf632f90 req-6d5025a0-ff68-48da-8261-39109646eb24 service nova] Releasing lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1163.451967] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465056, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.462600] env[61728]: DEBUG nova.network.neutron [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1163.496705] env[61728]: DEBUG oslo_vmware.api [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465057, 'name': PowerOffVM_Task, 'duration_secs': 0.246277} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.497844] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1163.497844] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1163.497844] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-88acac78-9b27-498e-9ab2-6389444c441c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.570868] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1163.571197] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1163.571413] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Deleting the datastore file [datastore1] 79782d2b-1b90-4f3f-a1ce-394e88a00546 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1163.571714] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63154184-47d9-4d7b-8bd1-1f42db79fdef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.581881] env[61728]: DEBUG oslo_vmware.api [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for the task: (returnval){ [ 1163.581881] env[61728]: value = "task-465059" [ 1163.581881] env[61728]: _type = "Task" [ 1163.581881] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.591146] env[61728]: DEBUG oslo_vmware.api [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465059, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.629442] env[61728]: DEBUG nova.network.neutron [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Updating instance_info_cache with network_info: [{"id": "704e9c0a-f4ff-48f7-a48f-ff4073e5929e", "address": "fa:16:3e:3c:6a:9e", "network": {"id": "c7098d8f-d31c-4d58-8ea6-b0043f3b21c8", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-251496334-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d34b224bdc2149288acb934783eb9954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap704e9c0a-f4", "ovs_interfaceid": "704e9c0a-f4ff-48f7-a48f-ff4073e5929e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1163.953797] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465056, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.418861} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.953797] env[61728]: INFO nova.virt.vmwareapi.ds_util [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_cf5a39d3-2fef-49d4-ac9f-06efde512f0b/OSTACK_IMG_cf5a39d3-2fef-49d4-ac9f-06efde512f0b.vmdk to [datastore1] devstack-image-cache_base/df2f6391-a88d-4bb9-ada0-dd01a6acec62/df2f6391-a88d-4bb9-ada0-dd01a6acec62.vmdk. [ 1163.954078] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Cleaning up location [datastore1] OSTACK_IMG_cf5a39d3-2fef-49d4-ac9f-06efde512f0b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1163.954928] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_cf5a39d3-2fef-49d4-ac9f-06efde512f0b {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1163.954928] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff73e70c-a83f-4a83-8d30-e6ea3e303deb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.962257] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1163.962257] env[61728]: value = "task-465060" [ 1163.962257] env[61728]: _type = "Task" [ 1163.962257] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.971353] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465060, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.102757] env[61728]: DEBUG oslo_vmware.api [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Task: {'id': task-465059, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.40382} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.103127] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.103332] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1164.104041] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1164.104041] env[61728]: INFO nova.compute.manager [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1164.104041] env[61728]: DEBUG oslo.service.loopingcall [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1164.104185] env[61728]: DEBUG nova.compute.manager [-] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1164.104270] env[61728]: DEBUG nova.network.neutron [-] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1164.132833] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Releasing lock "refresh_cache-76429588-fb4b-4215-b78c-75ce366989ed" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.133892] env[61728]: DEBUG nova.compute.manager [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Instance network_info: |[{"id": "704e9c0a-f4ff-48f7-a48f-ff4073e5929e", "address": "fa:16:3e:3c:6a:9e", "network": {"id": "c7098d8f-d31c-4d58-8ea6-b0043f3b21c8", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-251496334-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d34b224bdc2149288acb934783eb9954", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec903a9-d773-4d7c-a80c-c2533be346fb", "external-id": "nsx-vlan-transportzone-208", "segmentation_id": 208, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap704e9c0a-f4", "ovs_interfaceid": "704e9c0a-f4ff-48f7-a48f-ff4073e5929e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1164.134652] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:6a:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec903a9-d773-4d7c-a80c-c2533be346fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '704e9c0a-f4ff-48f7-a48f-ff4073e5929e', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1164.145060] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Creating folder: Project (d34b224bdc2149288acb934783eb9954). Parent ref: group-v121913. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1164.145378] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-60570f06-e8dc-4a9f-8698-81f30c62967a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.158114] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Created folder: Project (d34b224bdc2149288acb934783eb9954) in parent group-v121913. [ 1164.158344] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Creating folder: Instances. Parent ref: group-v122239. {{(pid=61728) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1164.158789] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-16ebf015-6899-4f44-aa09-392bdee99c0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.168920] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Created folder: Instances in parent group-v122239. [ 1164.169200] env[61728]: DEBUG oslo.service.loopingcall [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1164.169409] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1164.169630] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-298b7daa-a5eb-4f21-9d4f-a57a754c46a7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.190984] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1164.190984] env[61728]: value = "task-465063" [ 1164.190984] env[61728]: _type = "Task" [ 1164.190984] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.198455] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465063, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.368012] env[61728]: DEBUG nova.compute.manager [req-0fafed2c-b328-44fd-a151-20eaf2088f57 req-e8b904c4-5111-4521-bfa1-e081204ce625 service nova] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Received event network-vif-deleted-28c67e6e-1201-4072-8c7d-331a803bada0 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1164.368243] env[61728]: INFO nova.compute.manager [req-0fafed2c-b328-44fd-a151-20eaf2088f57 req-e8b904c4-5111-4521-bfa1-e081204ce625 service nova] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Neutron deleted interface 28c67e6e-1201-4072-8c7d-331a803bada0; detaching it from the instance and deleting it from the info cache [ 1164.368424] env[61728]: DEBUG nova.network.neutron [req-0fafed2c-b328-44fd-a151-20eaf2088f57 req-e8b904c4-5111-4521-bfa1-e081204ce625 service nova] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.471946] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465060, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.050022} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.472270] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.472372] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lock "[datastore1] devstack-image-cache_base/df2f6391-a88d-4bb9-ada0-dd01a6acec62/df2f6391-a88d-4bb9-ada0-dd01a6acec62.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.472587] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/df2f6391-a88d-4bb9-ada0-dd01a6acec62/df2f6391-a88d-4bb9-ada0-dd01a6acec62.vmdk to [datastore1] 532507fa-996d-45c7-bf70-f09c93be79ed/532507fa-996d-45c7-bf70-f09c93be79ed.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1164.472846] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5df04823-e637-4953-ae2f-52cfee620690 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.480103] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1164.480103] env[61728]: value = "task-465064" [ 1164.480103] env[61728]: _type = "Task" [ 1164.480103] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.488080] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465064, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.703258] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465063, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.773702] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "6218d586-0190-4a02-ae88-3f74329dc32c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.773979] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "6218d586-0190-4a02-ae88-3f74329dc32c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.840043] env[61728]: DEBUG nova.network.neutron [-] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.872543] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b7c9a14-0470-4b48-83ab-842efaa1bbb4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.882065] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8837b403-15a7-4fec-bbe9-8176ab3bd34d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.915498] env[61728]: DEBUG nova.compute.manager [req-0fafed2c-b328-44fd-a151-20eaf2088f57 req-e8b904c4-5111-4521-bfa1-e081204ce625 service nova] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Detach interface failed, port_id=28c67e6e-1201-4072-8c7d-331a803bada0, reason: Instance 79782d2b-1b90-4f3f-a1ce-394e88a00546 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1164.990331] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465064, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.202534] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465063, 'name': CreateVM_Task, 'duration_secs': 0.676526} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.202862] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1165.203516] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.203723] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.204055] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1165.204344] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a72b785-57e1-475b-a069-2e1a9ee5f052 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.209985] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Waiting for the task: (returnval){ [ 1165.209985] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5293a879-68df-b6a5-ad4e-3657dfd9511f" [ 1165.209985] env[61728]: _type = "Task" [ 1165.209985] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.218994] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5293a879-68df-b6a5-ad4e-3657dfd9511f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.276400] env[61728]: DEBUG nova.compute.manager [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1165.342873] env[61728]: INFO nova.compute.manager [-] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Took 1.24 seconds to deallocate network for instance. [ 1165.492505] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465064, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.721574] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5293a879-68df-b6a5-ad4e-3657dfd9511f, 'name': SearchDatastore_Task, 'duration_secs': 0.069669} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.721961] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1165.722226] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1165.722472] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.722627] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.722846] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1165.723917] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c5291504-05cc-4050-bcf9-feea2d232352 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.741308] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1165.742049] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1165.742491] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d64ea33f-89ce-4194-99be-7cba08594172 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.749948] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Waiting for the task: (returnval){ [ 1165.749948] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520b8500-83bc-1ca6-74be-840d0bc91388" [ 1165.749948] env[61728]: _type = "Task" [ 1165.749948] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.759485] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520b8500-83bc-1ca6-74be-840d0bc91388, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.803125] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.803422] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.805089] env[61728]: INFO nova.compute.claims [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1165.849512] env[61728]: DEBUG oslo_concurrency.lockutils [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.992728] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465064, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.261955] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]520b8500-83bc-1ca6-74be-840d0bc91388, 'name': SearchDatastore_Task, 'duration_secs': 0.076858} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.262769] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed0c5523-0bb6-4e5c-a1dc-ddd56d572427 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.269861] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Waiting for the task: (returnval){ [ 1166.269861] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5274fcf8-c45c-a3c4-34c8-3a9eb03390f9" [ 1166.269861] env[61728]: _type = "Task" [ 1166.269861] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.279262] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5274fcf8-c45c-a3c4-34c8-3a9eb03390f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.493812] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465064, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.779830] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5274fcf8-c45c-a3c4-34c8-3a9eb03390f9, 'name': SearchDatastore_Task, 'duration_secs': 0.078986} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.780119] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1166.780384] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 76429588-fb4b-4215-b78c-75ce366989ed/76429588-fb4b-4215-b78c-75ce366989ed.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1166.780649] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a6c5a7c3-a2b9-46bd-836c-3fde70651d15 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.786938] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Waiting for the task: (returnval){ [ 1166.786938] env[61728]: value = "task-465065" [ 1166.786938] env[61728]: _type = "Task" [ 1166.786938] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.794334] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465065, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.918026] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc83d27-2dac-4b04-b110-2a45c053ce21 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.925653] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ff6e9b-ddeb-40b1-8247-8f305d469929 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.958323] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7fd679-87a1-4ea2-9242-fdc162201b6d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.967361] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b81638c-9ca1-47c4-b6ed-325b1ee16cb0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.983832] env[61728]: DEBUG nova.compute.provider_tree [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1166.995686] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465064, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.28071} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.996701] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/df2f6391-a88d-4bb9-ada0-dd01a6acec62/df2f6391-a88d-4bb9-ada0-dd01a6acec62.vmdk to [datastore1] 532507fa-996d-45c7-bf70-f09c93be79ed/532507fa-996d-45c7-bf70-f09c93be79ed.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1166.997579] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a872f0ec-5234-4b5d-9b20-e95ef22b4a7f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.023140] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 532507fa-996d-45c7-bf70-f09c93be79ed/532507fa-996d-45c7-bf70-f09c93be79ed.vmdk or device None with type streamOptimized {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1167.023834] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4aae2141-0d3e-4bd6-95ad-51631fb1facd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.045130] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1167.045130] env[61728]: value = "task-465066" [ 1167.045130] env[61728]: _type = "Task" [ 1167.045130] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.056133] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465066, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.297608] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465065, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444125} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.298115] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 76429588-fb4b-4215-b78c-75ce366989ed/76429588-fb4b-4215-b78c-75ce366989ed.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1167.298115] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1167.298386] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f8d6f1e-89f5-4743-b942-7499043d490e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.305300] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Waiting for the task: (returnval){ [ 1167.305300] env[61728]: value = "task-465067" [ 1167.305300] env[61728]: _type = "Task" [ 1167.305300] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.314281] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465067, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.507752] env[61728]: ERROR nova.scheduler.client.report [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [req-76d63e53-e6eb-4b93-ad34-6069e65ec896] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-76d63e53-e6eb-4b93-ad34-6069e65ec896"}]} [ 1167.525104] env[61728]: DEBUG nova.scheduler.client.report [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1167.538362] env[61728]: DEBUG nova.scheduler.client.report [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1167.538599] env[61728]: DEBUG nova.compute.provider_tree [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1167.549501] env[61728]: DEBUG nova.scheduler.client.report [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1167.557308] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465066, 'name': ReconfigVM_Task, 'duration_secs': 0.377948} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.557577] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 532507fa-996d-45c7-bf70-f09c93be79ed/532507fa-996d-45c7-bf70-f09c93be79ed.vmdk or device None with type streamOptimized {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1167.558932] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_options': None, 'guest_format': None, 'encryption_secret_uuid': None, 'disk_bus': None, 'device_name': '/dev/sda', 'encryption_format': None, 'boot_index': 0, 'encrypted': False, 'size': 0, 'device_type': 'disk', 'image_id': '8b767102-1435-4827-a43b-8e2e25ec780b'}], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'c4b63838-2ce5-4b66-b34a-7bcd8dc37430', 'delete_on_termination': False, 'guest_format': None, 'disk_bus': None, 'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122234', 'volume_id': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'name': 'volume-ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '532507fa-996d-45c7-bf70-f09c93be79ed', 'attached_at': '', 'detached_at': '', 'volume_id': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'serial': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8'}, 'mount_device': '/dev/sdb', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=61728) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1167.559228] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1167.559385] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122234', 'volume_id': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'name': 'volume-ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '532507fa-996d-45c7-bf70-f09c93be79ed', 'attached_at': '', 'detached_at': '', 'volume_id': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'serial': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1167.560408] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad667b6-3d30-46ff-8fb6-ea9e3934fbb2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.577344] env[61728]: DEBUG nova.scheduler.client.report [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1167.579780] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9488502-81cd-42bf-9a86-de64e432bc25 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.609495] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] volume-ad2189db-dfcc-42fb-ab2a-15897b9ab1c8/volume-ad2189db-dfcc-42fb-ab2a-15897b9ab1c8.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1167.612120] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38d92aa7-03a1-40a5-be17-50555e1bb024 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.631489] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1167.631489] env[61728]: value = "task-465068" [ 1167.631489] env[61728]: _type = "Task" [ 1167.631489] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.642435] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465068, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.711453] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d53518-fc41-4a26-b932-ca64c6f48ef4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.719394] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2555586d-45aa-429d-9bdd-10c884240bea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.749609] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26782c97-ae94-46e3-af18-21c643ea491c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.757210] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c824fff0-3260-4c2f-b969-2fdcfcdfe925 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.770677] env[61728]: DEBUG nova.compute.provider_tree [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.814535] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465067, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069677} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.814763] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1167.815552] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af2a5a7-2dcf-4dda-9ab6-e37827f0a7ae {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.839370] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Reconfiguring VM instance instance-00000079 to attach disk [datastore1] 76429588-fb4b-4215-b78c-75ce366989ed/76429588-fb4b-4215-b78c-75ce366989ed.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1167.839596] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9779470-4872-4f73-a5aa-275b2444441c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.861726] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Waiting for the task: (returnval){ [ 1167.861726] env[61728]: value = "task-465069" [ 1167.861726] env[61728]: _type = "Task" [ 1167.861726] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.870047] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465069, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.143329] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465068, 'name': ReconfigVM_Task, 'duration_secs': 0.290187} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.143642] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Reconfigured VM instance instance-0000006f to attach disk [datastore1] volume-ad2189db-dfcc-42fb-ab2a-15897b9ab1c8/volume-ad2189db-dfcc-42fb-ab2a-15897b9ab1c8.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1168.148356] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2de63663-f9b9-49bc-9177-3df175301749 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.165107] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1168.165107] env[61728]: value = "task-465070" [ 1168.165107] env[61728]: _type = "Task" [ 1168.165107] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.175488] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465070, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.274286] env[61728]: DEBUG nova.scheduler.client.report [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1168.372982] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465069, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.677682] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465070, 'name': ReconfigVM_Task, 'duration_secs': 0.146726} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.678146] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122234', 'volume_id': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'name': 'volume-ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '532507fa-996d-45c7-bf70-f09c93be79ed', 'attached_at': '', 'detached_at': '', 'volume_id': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'serial': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1168.678948] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e93a248f-cd9d-42b4-97b8-166706dac0c7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.686235] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1168.686235] env[61728]: value = "task-465071" [ 1168.686235] env[61728]: _type = "Task" [ 1168.686235] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.694549] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465071, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.779727] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.976s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.780189] env[61728]: DEBUG nova.compute.manager [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1168.783111] env[61728]: DEBUG oslo_concurrency.lockutils [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.934s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.783251] env[61728]: DEBUG nova.objects.instance [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lazy-loading 'resources' on Instance uuid 79782d2b-1b90-4f3f-a1ce-394e88a00546 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.872096] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465069, 'name': ReconfigVM_Task, 'duration_secs': 0.756411} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.872385] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Reconfigured VM instance instance-00000079 to attach disk [datastore1] 76429588-fb4b-4215-b78c-75ce366989ed/76429588-fb4b-4215-b78c-75ce366989ed.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1168.873029] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ccfb21a-6308-46d6-906a-3c29033122cf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.879789] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Waiting for the task: (returnval){ [ 1168.879789] env[61728]: value = "task-465072" [ 1168.879789] env[61728]: _type = "Task" [ 1168.879789] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.888153] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465072, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.196189] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465071, 'name': Rename_Task, 'duration_secs': 0.142676} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.196478] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1169.196722] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec743481-453b-435b-8dab-8e87d9414ddf {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.203858] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1169.203858] env[61728]: value = "task-465073" [ 1169.203858] env[61728]: _type = "Task" [ 1169.203858] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.212796] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465073, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.286602] env[61728]: DEBUG nova.compute.utils [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1169.291274] env[61728]: DEBUG nova.compute.manager [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1169.291699] env[61728]: DEBUG nova.network.neutron [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1169.331397] env[61728]: DEBUG nova.policy [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3baf998a00e494ba33dab6a0a2a88e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '836843da5be34c649d9a48a83e658288', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1169.390892] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465072, 'name': Rename_Task, 'duration_secs': 0.139281} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.392188] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1169.392945] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7582b48-260a-4ee3-a1a3-9ca735fc36e4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.395422] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f94767e-67e5-4e33-93db-f9dc66110ae0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.402311] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa77bc80-4fc3-4248-8bb4-e2800e1343cb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.406637] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Waiting for the task: (returnval){ [ 1169.406637] env[61728]: value = "task-465074" [ 1169.406637] env[61728]: _type = "Task" [ 1169.406637] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.435771] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcfb7bc3-dabc-4370-b1ee-d83f73f34ca5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.441835] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465074, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.447316] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3e53313-3850-4e0a-955c-97f2d728095c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.462080] env[61728]: DEBUG nova.compute.provider_tree [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1169.630710] env[61728]: DEBUG nova.network.neutron [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Successfully created port: e30c4bdc-f4f7-4bf7-bf54-d980230430e5 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1169.714673] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465073, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.792443] env[61728]: DEBUG nova.compute.manager [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1169.918443] env[61728]: DEBUG oslo_vmware.api [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465074, 'name': PowerOnVM_Task, 'duration_secs': 0.495608} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.918840] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1169.919163] env[61728]: INFO nova.compute.manager [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Took 8.93 seconds to spawn the instance on the hypervisor. [ 1169.919498] env[61728]: DEBUG nova.compute.manager [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1169.920397] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7f1d58-40af-4410-917a-27afe3abf34a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.965879] env[61728]: DEBUG nova.scheduler.client.report [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1170.215695] env[61728]: DEBUG oslo_vmware.api [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465073, 'name': PowerOnVM_Task, 'duration_secs': 0.593149} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.215695] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1170.298013] env[61728]: INFO nova.virt.block_device [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Booting with volume 7bbf2464-ae28-4860-abbf-d14f03be449a at /dev/sda [ 1170.334895] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6fb11907-143a-496c-8f96-d220b797c95a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.341229] env[61728]: DEBUG nova.compute.manager [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1170.342186] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6094cb2e-58b9-4786-9ab7-9108a28cb020 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.349200] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7fce2e-7113-41db-8a30-0f34c247f484 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.381986] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8251092-fb00-4845-bb1a-c8df0f86f61d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.390413] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-364f2834-99cd-4874-9086-1c4bfe1ca37d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.419940] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85d17e9-eea3-499c-996c-68f5f22842bb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.426815] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43d006b-97d5-496e-930a-2a88e346cb15 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.437464] env[61728]: INFO nova.compute.manager [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Took 13.69 seconds to build instance. [ 1170.441516] env[61728]: DEBUG nova.virt.block_device [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updating existing volume attachment record: 0da3ce6e-9ad2-4601-920a-f60f464713fe {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1170.469577] env[61728]: DEBUG oslo_concurrency.lockutils [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.686s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.489940] env[61728]: INFO nova.scheduler.client.report [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Deleted allocations for instance 79782d2b-1b90-4f3f-a1ce-394e88a00546 [ 1170.871649] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5e8ac37f-4fd8-456a-a033-acc2b6af5089 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "532507fa-996d-45c7-bf70-f09c93be79ed" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 28.392s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.945817] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c337b8a5-eb03-4714-a74c-5b1e0b1bc48d tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Lock "76429588-fb4b-4215-b78c-75ce366989ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.204s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.998410] env[61728]: DEBUG oslo_concurrency.lockutils [None req-617d639b-d3f4-4792-a81b-0ae869743d4c tempest-ServerRescueTestJSON-843116773 tempest-ServerRescueTestJSON-843116773-project-member] Lock "79782d2b-1b90-4f3f-a1ce-394e88a00546" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.042s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.054865] env[61728]: DEBUG nova.compute.manager [req-95f47f94-dc24-42e0-8d60-c7805891bea0 req-e59dc45c-d77e-4823-8301-1b5fa1a4e549 service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Received event network-vif-plugged-e30c4bdc-f4f7-4bf7-bf54-d980230430e5 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1171.055079] env[61728]: DEBUG oslo_concurrency.lockutils [req-95f47f94-dc24-42e0-8d60-c7805891bea0 req-e59dc45c-d77e-4823-8301-1b5fa1a4e549 service nova] Acquiring lock "6218d586-0190-4a02-ae88-3f74329dc32c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.055305] env[61728]: DEBUG oslo_concurrency.lockutils [req-95f47f94-dc24-42e0-8d60-c7805891bea0 req-e59dc45c-d77e-4823-8301-1b5fa1a4e549 service nova] Lock "6218d586-0190-4a02-ae88-3f74329dc32c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.055476] env[61728]: DEBUG oslo_concurrency.lockutils [req-95f47f94-dc24-42e0-8d60-c7805891bea0 req-e59dc45c-d77e-4823-8301-1b5fa1a4e549 service nova] Lock "6218d586-0190-4a02-ae88-3f74329dc32c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.055650] env[61728]: DEBUG nova.compute.manager [req-95f47f94-dc24-42e0-8d60-c7805891bea0 req-e59dc45c-d77e-4823-8301-1b5fa1a4e549 service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] No waiting events found dispatching network-vif-plugged-e30c4bdc-f4f7-4bf7-bf54-d980230430e5 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1171.055822] env[61728]: WARNING nova.compute.manager [req-95f47f94-dc24-42e0-8d60-c7805891bea0 req-e59dc45c-d77e-4823-8301-1b5fa1a4e549 service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Received unexpected event network-vif-plugged-e30c4bdc-f4f7-4bf7-bf54-d980230430e5 for instance with vm_state building and task_state block_device_mapping. [ 1171.127960] env[61728]: DEBUG nova.network.neutron [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Successfully updated port: e30c4bdc-f4f7-4bf7-bf54-d980230430e5 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1171.495457] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Acquiring lock "76429588-fb4b-4215-b78c-75ce366989ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.495457] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Lock "76429588-fb4b-4215-b78c-75ce366989ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.495457] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Acquiring lock "76429588-fb4b-4215-b78c-75ce366989ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.495457] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Lock "76429588-fb4b-4215-b78c-75ce366989ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.495457] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Lock "76429588-fb4b-4215-b78c-75ce366989ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.496332] env[61728]: INFO nova.compute.manager [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Terminating instance [ 1171.498279] env[61728]: DEBUG nova.compute.manager [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1171.498607] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1171.499554] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df6a3f8-5a4d-47ae-aea5-6a03e1abe7d4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.507711] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1171.508111] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d64aa82-359c-4e20-8aa9-61ec6b662ad6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.516530] env[61728]: DEBUG oslo_vmware.api [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Waiting for the task: (returnval){ [ 1171.516530] env[61728]: value = "task-465075" [ 1171.516530] env[61728]: _type = "Task" [ 1171.516530] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.528681] env[61728]: DEBUG oslo_vmware.api [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465075, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.630908] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "refresh_cache-6218d586-0190-4a02-ae88-3f74329dc32c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1171.630988] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "refresh_cache-6218d586-0190-4a02-ae88-3f74329dc32c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.631221] env[61728]: DEBUG nova.network.neutron [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1172.032887] env[61728]: DEBUG oslo_vmware.api [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465075, 'name': PowerOffVM_Task, 'duration_secs': 0.199255} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.033206] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1172.033463] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1172.033735] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce515280-c9a8-4f64-b2d4-cb30949fefe8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.108369] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1172.108568] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1172.108799] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Deleting the datastore file [datastore1] 76429588-fb4b-4215-b78c-75ce366989ed {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1172.109191] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b4c1b456-e0f4-47af-a86c-36552d253a27 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.119988] env[61728]: DEBUG oslo_vmware.api [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Waiting for the task: (returnval){ [ 1172.119988] env[61728]: value = "task-465077" [ 1172.119988] env[61728]: _type = "Task" [ 1172.119988] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.127177] env[61728]: DEBUG oslo_vmware.api [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465077, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.178117] env[61728]: DEBUG nova.network.neutron [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1172.340749] env[61728]: DEBUG nova.network.neutron [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updating instance_info_cache with network_info: [{"id": "e30c4bdc-f4f7-4bf7-bf54-d980230430e5", "address": "fa:16:3e:fd:3b:b0", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape30c4bdc-f4", "ovs_interfaceid": "e30c4bdc-f4f7-4bf7-bf54-d980230430e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.531591] env[61728]: DEBUG nova.compute.manager [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1172.532251] env[61728]: DEBUG nova.virt.hardware [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1172.532487] env[61728]: DEBUG nova.virt.hardware [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1172.532650] env[61728]: DEBUG nova.virt.hardware [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1172.532918] env[61728]: DEBUG nova.virt.hardware [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1172.533115] env[61728]: DEBUG nova.virt.hardware [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1172.533307] env[61728]: DEBUG nova.virt.hardware [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1172.533545] env[61728]: DEBUG nova.virt.hardware [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1172.533717] env[61728]: DEBUG nova.virt.hardware [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1172.533896] env[61728]: DEBUG nova.virt.hardware [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1172.534088] env[61728]: DEBUG nova.virt.hardware [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1172.534281] env[61728]: DEBUG nova.virt.hardware [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1172.535197] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af1af62-fa6f-4455-bed7-d9fdebc0391d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.546111] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc680c67-e2aa-421b-859f-caa5b2ffbbd8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.627593] env[61728]: DEBUG oslo_vmware.api [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465077, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.843696] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "refresh_cache-6218d586-0190-4a02-ae88-3f74329dc32c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1172.843962] env[61728]: DEBUG nova.compute.manager [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Instance network_info: |[{"id": "e30c4bdc-f4f7-4bf7-bf54-d980230430e5", "address": "fa:16:3e:fd:3b:b0", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape30c4bdc-f4", "ovs_interfaceid": "e30c4bdc-f4f7-4bf7-bf54-d980230430e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1172.844506] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:3b:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9732690c-bdcf-4e6f-9a32-42c196333eb8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e30c4bdc-f4f7-4bf7-bf54-d980230430e5', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1172.851980] env[61728]: DEBUG oslo.service.loopingcall [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1172.852241] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1172.852470] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-43d9c15f-777d-4c90-9e1c-35f515153cb5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.873194] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1172.873194] env[61728]: value = "task-465078" [ 1172.873194] env[61728]: _type = "Task" [ 1172.873194] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.881438] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465078, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.083180] env[61728]: DEBUG nova.compute.manager [req-cfdc9ba3-e726-4459-b2c1-7953722c8a2c req-023e5258-8924-41aa-8e06-fc07d48de7ee service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Received event network-changed-e30c4bdc-f4f7-4bf7-bf54-d980230430e5 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1173.083408] env[61728]: DEBUG nova.compute.manager [req-cfdc9ba3-e726-4459-b2c1-7953722c8a2c req-023e5258-8924-41aa-8e06-fc07d48de7ee service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Refreshing instance network info cache due to event network-changed-e30c4bdc-f4f7-4bf7-bf54-d980230430e5. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1173.083634] env[61728]: DEBUG oslo_concurrency.lockutils [req-cfdc9ba3-e726-4459-b2c1-7953722c8a2c req-023e5258-8924-41aa-8e06-fc07d48de7ee service nova] Acquiring lock "refresh_cache-6218d586-0190-4a02-ae88-3f74329dc32c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1173.083784] env[61728]: DEBUG oslo_concurrency.lockutils [req-cfdc9ba3-e726-4459-b2c1-7953722c8a2c req-023e5258-8924-41aa-8e06-fc07d48de7ee service nova] Acquired lock "refresh_cache-6218d586-0190-4a02-ae88-3f74329dc32c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1173.083949] env[61728]: DEBUG nova.network.neutron [req-cfdc9ba3-e726-4459-b2c1-7953722c8a2c req-023e5258-8924-41aa-8e06-fc07d48de7ee service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Refreshing network info cache for port e30c4bdc-f4f7-4bf7-bf54-d980230430e5 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1173.128986] env[61728]: DEBUG oslo_vmware.api [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Task: {'id': task-465077, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.811295} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.129228] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1173.129448] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1173.129641] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1173.129862] env[61728]: INFO nova.compute.manager [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Took 1.63 seconds to destroy the instance on the hypervisor. [ 1173.130144] env[61728]: DEBUG oslo.service.loopingcall [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1173.130416] env[61728]: DEBUG nova.compute.manager [-] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1173.130488] env[61728]: DEBUG nova.network.neutron [-] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1173.385522] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465078, 'name': CreateVM_Task, 'duration_secs': 0.366547} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.386865] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1173.387579] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': '0da3ce6e-9ad2-4601-920a-f60f464713fe', 'delete_on_termination': True, 'guest_format': None, 'disk_bus': None, 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122235', 'volume_id': '7bbf2464-ae28-4860-abbf-d14f03be449a', 'name': 'volume-7bbf2464-ae28-4860-abbf-d14f03be449a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6218d586-0190-4a02-ae88-3f74329dc32c', 'attached_at': '', 'detached_at': '', 'volume_id': '7bbf2464-ae28-4860-abbf-d14f03be449a', 'serial': '7bbf2464-ae28-4860-abbf-d14f03be449a'}, 'mount_device': '/dev/sda', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=61728) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1173.387802] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Root volume attach. Driver type: vmdk {{(pid=61728) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1173.388907] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b011f9-5ad8-42e8-8d44-cb735dce7f8c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.397647] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f0425cc-a6a8-4ee4-b471-e979c28ed1bb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.405015] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d2ef26-1a0c-43b8-ad8d-8d27e56b57d1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.411088] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-f6079cae-b63e-4e7a-b0db-3086da89dd58 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.419156] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1173.419156] env[61728]: value = "task-465079" [ 1173.419156] env[61728]: _type = "Task" [ 1173.419156] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.426894] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465079, 'name': RelocateVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.778247] env[61728]: DEBUG nova.network.neutron [req-cfdc9ba3-e726-4459-b2c1-7953722c8a2c req-023e5258-8924-41aa-8e06-fc07d48de7ee service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updated VIF entry in instance network info cache for port e30c4bdc-f4f7-4bf7-bf54-d980230430e5. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1173.778630] env[61728]: DEBUG nova.network.neutron [req-cfdc9ba3-e726-4459-b2c1-7953722c8a2c req-023e5258-8924-41aa-8e06-fc07d48de7ee service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updating instance_info_cache with network_info: [{"id": "e30c4bdc-f4f7-4bf7-bf54-d980230430e5", "address": "fa:16:3e:fd:3b:b0", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape30c4bdc-f4", "ovs_interfaceid": "e30c4bdc-f4f7-4bf7-bf54-d980230430e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1173.889340] env[61728]: DEBUG nova.network.neutron [-] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1173.929547] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465079, 'name': RelocateVM_Task, 'duration_secs': 0.419068} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.929974] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1173.930151] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122235', 'volume_id': '7bbf2464-ae28-4860-abbf-d14f03be449a', 'name': 'volume-7bbf2464-ae28-4860-abbf-d14f03be449a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6218d586-0190-4a02-ae88-3f74329dc32c', 'attached_at': '', 'detached_at': '', 'volume_id': '7bbf2464-ae28-4860-abbf-d14f03be449a', 'serial': '7bbf2464-ae28-4860-abbf-d14f03be449a'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1173.930927] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889cade7-5614-456c-8459-3c575e559fd9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.948700] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c6cef4-e263-403a-83b5-3cdadfb25121 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.971094] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Reconfiguring VM instance instance-0000007a to attach disk [datastore1] volume-7bbf2464-ae28-4860-abbf-d14f03be449a/volume-7bbf2464-ae28-4860-abbf-d14f03be449a.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1173.971740] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13d8644f-74a3-4e91-a949-385d708a58c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.991263] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1173.991263] env[61728]: value = "task-465080" [ 1173.991263] env[61728]: _type = "Task" [ 1173.991263] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.999815] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465080, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.281497] env[61728]: DEBUG oslo_concurrency.lockutils [req-cfdc9ba3-e726-4459-b2c1-7953722c8a2c req-023e5258-8924-41aa-8e06-fc07d48de7ee service nova] Releasing lock "refresh_cache-6218d586-0190-4a02-ae88-3f74329dc32c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1174.392481] env[61728]: INFO nova.compute.manager [-] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Took 1.26 seconds to deallocate network for instance. [ 1174.501735] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465080, 'name': ReconfigVM_Task, 'duration_secs': 0.344861} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.502010] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Reconfigured VM instance instance-0000007a to attach disk [datastore1] volume-7bbf2464-ae28-4860-abbf-d14f03be449a/volume-7bbf2464-ae28-4860-abbf-d14f03be449a.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1174.506709] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9dee1a8e-40f7-4587-91ea-5580bf366c02 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.522228] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1174.522228] env[61728]: value = "task-465081" [ 1174.522228] env[61728]: _type = "Task" [ 1174.522228] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.530288] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465081, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.899886] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.900423] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.900501] env[61728]: DEBUG nova.objects.instance [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Lazy-loading 'resources' on Instance uuid 76429588-fb4b-4215-b78c-75ce366989ed {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1175.033017] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465081, 'name': ReconfigVM_Task, 'duration_secs': 0.119967} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.033502] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122235', 'volume_id': '7bbf2464-ae28-4860-abbf-d14f03be449a', 'name': 'volume-7bbf2464-ae28-4860-abbf-d14f03be449a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6218d586-0190-4a02-ae88-3f74329dc32c', 'attached_at': '', 'detached_at': '', 'volume_id': '7bbf2464-ae28-4860-abbf-d14f03be449a', 'serial': '7bbf2464-ae28-4860-abbf-d14f03be449a'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1175.033970] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-71a1c4ff-e4fa-43f3-8585-d930fffeede9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.040917] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1175.040917] env[61728]: value = "task-465082" [ 1175.040917] env[61728]: _type = "Task" [ 1175.040917] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.048741] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465082, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.107125] env[61728]: DEBUG nova.compute.manager [req-0d74ca2a-56cc-447b-a0b7-a2b64e0f0387 req-f1620589-e8cd-4646-8f14-efff6dce9d31 service nova] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Received event network-vif-deleted-704e9c0a-f4ff-48f7-a48f-ff4073e5929e {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1175.492082] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1785934c-9c7e-4411-ae6c-ff4e401709c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.499681] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387d20e3-7ff9-4a67-a86f-786e95b70f1b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.530859] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b395fe1a-025f-4580-938d-8b61abe76b25 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.538916] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2efd6fed-4a12-4144-bc73-5bb27ac3419b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.555457] env[61728]: DEBUG nova.compute.provider_tree [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1175.558950] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465082, 'name': Rename_Task, 'duration_secs': 0.149524} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.559417] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1175.559640] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cce9f060-8c20-4516-a27e-bea08556cc5a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.565842] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1175.565842] env[61728]: value = "task-465083" [ 1175.565842] env[61728]: _type = "Task" [ 1175.565842] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.573517] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465083, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.061367] env[61728]: DEBUG nova.scheduler.client.report [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1176.076605] env[61728]: DEBUG oslo_vmware.api [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465083, 'name': PowerOnVM_Task, 'duration_secs': 0.431788} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.077479] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1176.077694] env[61728]: INFO nova.compute.manager [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Took 3.55 seconds to spawn the instance on the hypervisor. [ 1176.077879] env[61728]: DEBUG nova.compute.manager [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1176.078659] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52938e9b-0909-4120-b2ae-a03856883641 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.566431] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.666s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.583942] env[61728]: INFO nova.scheduler.client.report [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Deleted allocations for instance 76429588-fb4b-4215-b78c-75ce366989ed [ 1176.612051] env[61728]: INFO nova.compute.manager [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Took 10.83 seconds to build instance. [ 1177.091129] env[61728]: DEBUG oslo_concurrency.lockutils [None req-7dda4349-d00b-4882-8d29-b303b8cd8dcb tempest-ServerAddressesTestJSON-1067598512 tempest-ServerAddressesTestJSON-1067598512-project-member] Lock "76429588-fb4b-4215-b78c-75ce366989ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.598s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.113711] env[61728]: DEBUG oslo_concurrency.lockutils [None req-e435286c-4290-4260-a03e-1d58aaa262a1 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "6218d586-0190-4a02-ae88-3f74329dc32c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.340s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.944876] env[61728]: DEBUG nova.compute.manager [req-6dd3db25-5392-46b2-865a-279464bc4b1d req-17cbf3d0-6204-4a05-a15a-ead71d3afaee service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Received event network-changed-51b6c967-37bc-4156-a23f-91040bb1c4a2 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1177.945586] env[61728]: DEBUG nova.compute.manager [req-6dd3db25-5392-46b2-865a-279464bc4b1d req-17cbf3d0-6204-4a05-a15a-ead71d3afaee service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Refreshing instance network info cache due to event network-changed-51b6c967-37bc-4156-a23f-91040bb1c4a2. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1177.945875] env[61728]: DEBUG oslo_concurrency.lockutils [req-6dd3db25-5392-46b2-865a-279464bc4b1d req-17cbf3d0-6204-4a05-a15a-ead71d3afaee service nova] Acquiring lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1177.946055] env[61728]: DEBUG oslo_concurrency.lockutils [req-6dd3db25-5392-46b2-865a-279464bc4b1d req-17cbf3d0-6204-4a05-a15a-ead71d3afaee service nova] Acquired lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.946235] env[61728]: DEBUG nova.network.neutron [req-6dd3db25-5392-46b2-865a-279464bc4b1d req-17cbf3d0-6204-4a05-a15a-ead71d3afaee service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Refreshing network info cache for port 51b6c967-37bc-4156-a23f-91040bb1c4a2 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1178.852269] env[61728]: DEBUG nova.network.neutron [req-6dd3db25-5392-46b2-865a-279464bc4b1d req-17cbf3d0-6204-4a05-a15a-ead71d3afaee service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Updated VIF entry in instance network info cache for port 51b6c967-37bc-4156-a23f-91040bb1c4a2. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1178.852661] env[61728]: DEBUG nova.network.neutron [req-6dd3db25-5392-46b2-865a-279464bc4b1d req-17cbf3d0-6204-4a05-a15a-ead71d3afaee service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Updating instance_info_cache with network_info: [{"id": "51b6c967-37bc-4156-a23f-91040bb1c4a2", "address": "fa:16:3e:7f:f1:5d", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b6c967-37", "ovs_interfaceid": "51b6c967-37bc-4156-a23f-91040bb1c4a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1179.356973] env[61728]: DEBUG oslo_concurrency.lockutils [req-6dd3db25-5392-46b2-865a-279464bc4b1d req-17cbf3d0-6204-4a05-a15a-ead71d3afaee service nova] Releasing lock "refresh_cache-b1fb2e3f-234a-44c7-983e-d4441b4e3cef" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.756331] env[61728]: DEBUG nova.compute.manager [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Stashing vm_state: active {{(pid=61728) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1179.759049] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1179.759420] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1179.970331] env[61728]: DEBUG nova.compute.manager [req-cb76b4ff-c035-4c83-9c46-1855ada36a9c req-0d5fd999-42d8-48fc-be20-0dc84211ed80 service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Received event network-changed-e30c4bdc-f4f7-4bf7-bf54-d980230430e5 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1179.970689] env[61728]: DEBUG nova.compute.manager [req-cb76b4ff-c035-4c83-9c46-1855ada36a9c req-0d5fd999-42d8-48fc-be20-0dc84211ed80 service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Refreshing instance network info cache due to event network-changed-e30c4bdc-f4f7-4bf7-bf54-d980230430e5. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1179.970989] env[61728]: DEBUG oslo_concurrency.lockutils [req-cb76b4ff-c035-4c83-9c46-1855ada36a9c req-0d5fd999-42d8-48fc-be20-0dc84211ed80 service nova] Acquiring lock "refresh_cache-6218d586-0190-4a02-ae88-3f74329dc32c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.971213] env[61728]: DEBUG oslo_concurrency.lockutils [req-cb76b4ff-c035-4c83-9c46-1855ada36a9c req-0d5fd999-42d8-48fc-be20-0dc84211ed80 service nova] Acquired lock "refresh_cache-6218d586-0190-4a02-ae88-3f74329dc32c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.971457] env[61728]: DEBUG nova.network.neutron [req-cb76b4ff-c035-4c83-9c46-1855ada36a9c req-0d5fd999-42d8-48fc-be20-0dc84211ed80 service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Refreshing network info cache for port e30c4bdc-f4f7-4bf7-bf54-d980230430e5 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1180.266227] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.266428] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Starting heal instance info cache {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1180.280722] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1180.280988] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1180.683972] env[61728]: DEBUG nova.network.neutron [req-cb76b4ff-c035-4c83-9c46-1855ada36a9c req-0d5fd999-42d8-48fc-be20-0dc84211ed80 service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updated VIF entry in instance network info cache for port e30c4bdc-f4f7-4bf7-bf54-d980230430e5. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1180.684391] env[61728]: DEBUG nova.network.neutron [req-cb76b4ff-c035-4c83-9c46-1855ada36a9c req-0d5fd999-42d8-48fc-be20-0dc84211ed80 service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updating instance_info_cache with network_info: [{"id": "e30c4bdc-f4f7-4bf7-bf54-d980230430e5", "address": "fa:16:3e:fd:3b:b0", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape30c4bdc-f4", "ovs_interfaceid": "e30c4bdc-f4f7-4bf7-bf54-d980230430e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.785951] env[61728]: INFO nova.compute.claims [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1181.187441] env[61728]: DEBUG oslo_concurrency.lockutils [req-cb76b4ff-c035-4c83-9c46-1855ada36a9c req-0d5fd999-42d8-48fc-be20-0dc84211ed80 service nova] Releasing lock "refresh_cache-6218d586-0190-4a02-ae88-3f74329dc32c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.292134] env[61728]: INFO nova.compute.resource_tracker [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updating resource usage from migration 25344c24-1fd6-4dd8-ad44-3ef01d41d85d [ 1181.386057] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e2a6eb-9cf1-4f0a-abd5-8b4c6b7e66ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.396354] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd12953-89aa-42d0-a06f-3105ebb48f8b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.440833] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f97c0a3c-db5c-439b-8e1b-8184665e30ec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.448289] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-025cd0b4-6d68-4aac-8cf4-6e1bad49d576 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.461686] env[61728]: DEBUG nova.compute.provider_tree [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1181.965063] env[61728]: DEBUG nova.scheduler.client.report [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1182.470072] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.189s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1182.470448] env[61728]: INFO nova.compute.manager [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Migrating [ 1182.985781] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "refresh_cache-6218d586-0190-4a02-ae88-3f74329dc32c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.985961] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquired lock "refresh_cache-6218d586-0190-4a02-ae88-3f74329dc32c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.986134] env[61728]: DEBUG nova.network.neutron [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1183.686965] env[61728]: DEBUG nova.network.neutron [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updating instance_info_cache with network_info: [{"id": "e30c4bdc-f4f7-4bf7-bf54-d980230430e5", "address": "fa:16:3e:fd:3b:b0", "network": {"id": "816c2811-2114-4893-b93d-957134e4471a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1972112539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "836843da5be34c649d9a48a83e658288", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape30c4bdc-f4", "ovs_interfaceid": "e30c4bdc-f4f7-4bf7-bf54-d980230430e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1183.769482] env[61728]: DEBUG oslo_concurrency.lockutils [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.769822] env[61728]: DEBUG oslo_concurrency.lockutils [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.770134] env[61728]: DEBUG oslo_concurrency.lockutils [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "d55beb02-00fc-4df6-a239-9e5d776a7c18-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.770392] env[61728]: DEBUG oslo_concurrency.lockutils [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.770630] env[61728]: DEBUG oslo_concurrency.lockutils [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.772981] env[61728]: INFO nova.compute.manager [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Terminating instance [ 1183.774895] env[61728]: DEBUG nova.compute.manager [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1183.775153] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1183.776051] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1e2a9c-3e94-4c92-9400-42ad03e35a4e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.786185] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1183.786620] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e061ed34-1073-4f8b-b555-4355aec34df3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.793920] env[61728]: DEBUG oslo_vmware.api [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1183.793920] env[61728]: value = "task-465084" [ 1183.793920] env[61728]: _type = "Task" [ 1183.793920] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.803584] env[61728]: DEBUG oslo_vmware.api [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465084, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.189960] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Releasing lock "refresh_cache-6218d586-0190-4a02-ae88-3f74329dc32c" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.287073] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Didn't find any instances for network info cache update. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1184.287342] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1184.287507] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1184.287657] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1184.287806] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1184.287948] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1184.288109] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1184.288244] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61728) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1184.288389] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1184.304861] env[61728]: DEBUG oslo_vmware.api [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465084, 'name': PowerOffVM_Task, 'duration_secs': 0.227886} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.305146] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1184.305354] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1184.305609] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d36d617-1fea-44d3-80ce-d65b7e230be7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.370282] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1184.370523] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1184.370717] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Deleting the datastore file [datastore1] d55beb02-00fc-4df6-a239-9e5d776a7c18 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1184.370993] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5dab06ef-194f-4f02-9889-5aabfd2f72df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.377562] env[61728]: DEBUG oslo_vmware.api [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1184.377562] env[61728]: value = "task-465086" [ 1184.377562] env[61728]: _type = "Task" [ 1184.377562] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.385236] env[61728]: DEBUG oslo_vmware.api [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465086, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.791437] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.791844] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.791844] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.792019] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61728) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1184.792891] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20293528-d8d5-4be1-8209-1f19dfffbbb4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.801163] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8072ebf2-0f58-4b5c-959f-bbfaff3a05bd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.814831] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f139996-7ea2-4eeb-b537-37e6b8df5683 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.822856] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e840f801-51b8-4086-b051-bace3a86fc44 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.851321] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180921MB free_disk=115GB free_vcpus=48 pci_devices=None {{(pid=61728) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1184.851563] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.851742] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.888304] env[61728]: DEBUG oslo_vmware.api [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465086, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134216} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.888645] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1184.888733] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1184.888915] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1184.889113] env[61728]: INFO nova.compute.manager [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1184.889364] env[61728]: DEBUG oslo.service.loopingcall [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1184.889560] env[61728]: DEBUG nova.compute.manager [-] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1184.889658] env[61728]: DEBUG nova.network.neutron [-] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1185.310357] env[61728]: DEBUG nova.compute.manager [req-3529d3a3-fafd-47f6-91e5-5e99a7cd6423 req-59e5042e-5fb3-4030-a789-8cf07f864aaf service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Received event network-vif-deleted-4155b1a7-975f-404c-8374-47665c8046ca {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1185.310521] env[61728]: INFO nova.compute.manager [req-3529d3a3-fafd-47f6-91e5-5e99a7cd6423 req-59e5042e-5fb3-4030-a789-8cf07f864aaf service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Neutron deleted interface 4155b1a7-975f-404c-8374-47665c8046ca; detaching it from the instance and deleting it from the info cache [ 1185.310706] env[61728]: DEBUG nova.network.neutron [req-3529d3a3-fafd-47f6-91e5-5e99a7cd6423 req-59e5042e-5fb3-4030-a789-8cf07f864aaf service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.705328] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ff1b3c-9365-4f68-95f0-9e7baef0bcb1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.724265] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updating instance '6218d586-0190-4a02-ae88-3f74329dc32c' progress to 0 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1185.790524] env[61728]: DEBUG nova.network.neutron [-] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.813642] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-73479caa-d361-48e7-851f-ff8f7b939798 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.823928] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f75932-b1ce-4be6-9e18-2d782da60e26 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.853613] env[61728]: DEBUG nova.compute.manager [req-3529d3a3-fafd-47f6-91e5-5e99a7cd6423 req-59e5042e-5fb3-4030-a789-8cf07f864aaf service nova] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Detach interface failed, port_id=4155b1a7-975f-404c-8374-47665c8046ca, reason: Instance d55beb02-00fc-4df6-a239-9e5d776a7c18 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1185.859848] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Applying migration context for instance 6218d586-0190-4a02-ae88-3f74329dc32c as it has an incoming, in-progress migration 25344c24-1fd6-4dd8-ad44-3ef01d41d85d. Migration status is migrating {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1185.860831] env[61728]: INFO nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updating resource usage from migration 25344c24-1fd6-4dd8-ad44-3ef01d41d85d [ 1185.879730] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance b1fb2e3f-234a-44c7-983e-d4441b4e3cef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1185.880008] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance d55beb02-00fc-4df6-a239-9e5d776a7c18 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1185.880179] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 2db96467-5fb4-49f4-954b-dbdd21e46232 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1185.880311] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 532507fa-996d-45c7-bf70-f09c93be79ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1185.880429] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance fb1195b2-1ab5-47d7-be86-4b61e47b4e29 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1185.880552] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Migration 25344c24-1fd6-4dd8-ad44-3ef01d41d85d is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1185.880959] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 6218d586-0190-4a02-ae88-3f74329dc32c actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1185.880959] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1185.881116] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1920MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1185.996575] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed1a564-e3a4-4db6-9d89-8283007b25c9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.008766] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59bf09ef-47e5-4b78-83f2-3710aecc9d4b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.056062] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f78558-6b2c-472e-aa7f-09467d1b5e6b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.065167] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2a1f6ee-e127-4509-8d56-e5362a0a0b30 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.080773] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1186.229751] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1186.230066] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28e511bc-2c38-413c-8004-27b3423561e9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.237736] env[61728]: DEBUG oslo_vmware.api [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1186.237736] env[61728]: value = "task-465087" [ 1186.237736] env[61728]: _type = "Task" [ 1186.237736] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.246870] env[61728]: DEBUG oslo_vmware.api [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465087, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.296106] env[61728]: INFO nova.compute.manager [-] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Took 1.41 seconds to deallocate network for instance. [ 1186.600397] env[61728]: ERROR nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [req-afce32ad-7f2a-47fc-a57d-0231e0aacfa9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e7ceb92f-072b-409e-b888-6fe0676b32f1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-afce32ad-7f2a-47fc-a57d-0231e0aacfa9"}]} [ 1186.615924] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Refreshing inventories for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1186.630343] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Updating ProviderTree inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1186.630544] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 114, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1186.640579] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Refreshing aggregate associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, aggregates: None {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1186.656718] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Refreshing trait associations for resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61728) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1186.731844] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68edae29-96ac-46af-b081-1290fdaad66f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.742396] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83457cec-ed58-466d-a1c3-21e380af8ecc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.750462] env[61728]: DEBUG oslo_vmware.api [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465087, 'name': PowerOffVM_Task, 'duration_secs': 0.169} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.774649] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1186.774885] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updating instance '6218d586-0190-4a02-ae88-3f74329dc32c' progress to 17 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1186.779088] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79b2207-8fd7-49eb-a1e2-93a9879fbd3c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.787064] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc417574-a3b8-4339-a760-bdf0b230d3bb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.802344] env[61728]: DEBUG oslo_concurrency.lockutils [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.802777] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1187.281421] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1187.281886] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1187.281886] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1187.282054] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1187.282216] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1187.282374] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1187.282584] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1187.282751] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1187.282975] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1187.283169] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1187.283349] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1187.288311] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7ad010d-33dd-41cf-91d0-a2a0d54bdf92 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.307792] env[61728]: DEBUG oslo_vmware.api [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1187.307792] env[61728]: value = "task-465088" [ 1187.307792] env[61728]: _type = "Task" [ 1187.307792] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.317017] env[61728]: DEBUG oslo_vmware.api [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465088, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.335007] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Updated inventory for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with generation 193 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1187.335267] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Updating resource provider e7ceb92f-072b-409e-b888-6fe0676b32f1 generation from 193 to 194 during operation: update_inventory {{(pid=61728) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1187.335421] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Updating inventory in ProviderTree for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1187.817679] env[61728]: DEBUG oslo_vmware.api [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465088, 'name': ReconfigVM_Task, 'duration_secs': 0.21876} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.818013] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updating instance '6218d586-0190-4a02-ae88-3f74329dc32c' progress to 33 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1187.840361] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1187.840571] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.989s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.840765] env[61728]: DEBUG oslo_concurrency.lockutils [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.039s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.840988] env[61728]: DEBUG nova.objects.instance [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lazy-loading 'resources' on Instance uuid d55beb02-00fc-4df6-a239-9e5d776a7c18 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.324322] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1188.324694] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1188.324747] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1188.324937] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1188.325103] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1188.325261] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1188.325486] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1188.325655] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1188.325833] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1188.326013] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1188.326206] env[61728]: DEBUG nova.virt.hardware [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1188.425869] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f580f6-9413-45e6-a132-e904379716f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.433975] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557125f2-bc0a-4c0a-bfdd-7889c75b6ba3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.464547] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-980ebc08-e7f7-4244-9c25-b8f604d9b3f1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.471645] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c45cf3-4cff-4447-9312-efd1cc754550 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.484653] env[61728]: DEBUG nova.compute.provider_tree [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1188.834708] env[61728]: ERROR nova.compute.manager [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Setting instance vm_state to ERROR: AttributeError: 'NoneType' object has no attribute 'key' [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Traceback (most recent call last): [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] yield [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] disk_info = self.driver.migrate_disk_and_power_off( [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 263, in migrate_disk_and_power_off [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] return self._vmops.migrate_disk_and_power_off(context, instance, [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] disk_key = device.key [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] AttributeError: 'NoneType' object has no attribute 'key' [ 1188.834708] env[61728]: ERROR nova.compute.manager [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] [ 1188.987771] env[61728]: DEBUG nova.scheduler.client.report [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1189.352771] env[61728]: INFO nova.compute.manager [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Swapping old allocation on dict_keys(['e7ceb92f-072b-409e-b888-6fe0676b32f1']) held by migration 25344c24-1fd6-4dd8-ad44-3ef01d41d85d for instance [ 1189.373532] env[61728]: DEBUG nova.scheduler.client.report [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Overwriting current allocation {'allocations': {'e7ceb92f-072b-409e-b888-6fe0676b32f1': {'resources': {'VCPU': 1, 'MEMORY_MB': 256}, 'generation': 194}}, 'project_id': '836843da5be34c649d9a48a83e658288', 'user_id': 'f3baf998a00e494ba33dab6a0a2a88e4', 'consumer_generation': 1} on consumer 6218d586-0190-4a02-ae88-3f74329dc32c {{(pid=61728) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1189.492572] env[61728]: DEBUG oslo_concurrency.lockutils [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.652s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.509979] env[61728]: INFO nova.scheduler.client.report [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Deleted allocations for instance d55beb02-00fc-4df6-a239-9e5d776a7c18 [ 1190.019584] env[61728]: DEBUG oslo_concurrency.lockutils [None req-980e2130-2b93-42fd-91d5-151a2469dae5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "d55beb02-00fc-4df6-a239-9e5d776a7c18" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.250s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.363173] env[61728]: DEBUG oslo_concurrency.lockutils [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "6218d586-0190-4a02-ae88-3f74329dc32c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.363173] env[61728]: DEBUG oslo_concurrency.lockutils [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "6218d586-0190-4a02-ae88-3f74329dc32c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.363537] env[61728]: DEBUG oslo_concurrency.lockutils [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "6218d586-0190-4a02-ae88-3f74329dc32c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.363586] env[61728]: DEBUG oslo_concurrency.lockutils [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "6218d586-0190-4a02-ae88-3f74329dc32c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.363752] env[61728]: DEBUG oslo_concurrency.lockutils [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "6218d586-0190-4a02-ae88-3f74329dc32c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.365930] env[61728]: INFO nova.compute.manager [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Terminating instance [ 1190.368044] env[61728]: DEBUG nova.compute.manager [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1190.368147] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1190.368435] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2fcdd72f-47ec-4a25-b536-ccb9aef90a0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.377415] env[61728]: DEBUG oslo_vmware.api [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1190.377415] env[61728]: value = "task-465089" [ 1190.377415] env[61728]: _type = "Task" [ 1190.377415] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.388317] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1190.388696] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1190.389199] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122235', 'volume_id': '7bbf2464-ae28-4860-abbf-d14f03be449a', 'name': 'volume-7bbf2464-ae28-4860-abbf-d14f03be449a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6218d586-0190-4a02-ae88-3f74329dc32c', 'attached_at': '', 'detached_at': '', 'volume_id': '7bbf2464-ae28-4860-abbf-d14f03be449a', 'serial': '7bbf2464-ae28-4860-abbf-d14f03be449a'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1190.389631] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf93cc0b-43b2-456b-b01d-79410d85d08a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.408629] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e93f3bb5-7430-4e9d-aeab-3b1bd0936f8e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.415571] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a2b0400-5459-47a1-97c7-8b92f5dcc5c3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.435880] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db50c20b-9bc3-42b7-a7b2-1e943a0a4a36 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.451008] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] The volume has not been displaced from its original location: [datastore1] volume-7bbf2464-ae28-4860-abbf-d14f03be449a/volume-7bbf2464-ae28-4860-abbf-d14f03be449a.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1190.456343] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Reconfiguring VM instance instance-0000007a to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1190.456654] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7eb3ab6c-b945-44ae-b787-2177ba55ba62 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.475641] env[61728]: DEBUG oslo_vmware.api [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1190.475641] env[61728]: value = "task-465090" [ 1190.475641] env[61728]: _type = "Task" [ 1190.475641] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.484391] env[61728]: DEBUG oslo_vmware.api [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465090, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.937231] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.937738] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.987639] env[61728]: DEBUG oslo_vmware.api [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465090, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.020066] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ffc24f-160c-4724-ac4a-5e55b5e3db66 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.028644] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33045147-4549-4852-8a0a-f0144e98f438 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.066672] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5351b36-fc06-4622-8236-af53c85edc47 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.076888] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-710d0580-deba-451f-aa01-3fb94c716481 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.093732] env[61728]: DEBUG nova.compute.provider_tree [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1191.487744] env[61728]: DEBUG oslo_vmware.api [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465090, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.599213] env[61728]: DEBUG nova.scheduler.client.report [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1191.674974] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.675192] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.744876] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "6218d586-0190-4a02-ae88-3f74329dc32c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.988113] env[61728]: DEBUG oslo_vmware.api [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465090, 'name': ReconfigVM_Task, 'duration_secs': 1.192166} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.988378] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Reconfigured VM instance instance-0000007a to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1191.993181] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f72d966e-c229-47c6-983c-f8904de82ddc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.010155] env[61728]: DEBUG oslo_vmware.api [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1192.010155] env[61728]: value = "task-465091" [ 1192.010155] env[61728]: _type = "Task" [ 1192.010155] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.019109] env[61728]: DEBUG oslo_vmware.api [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465091, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.105287] env[61728]: DEBUG oslo_concurrency.lockutils [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.168s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1192.105539] env[61728]: INFO nova.compute.manager [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Successfully reverted task state from resize_migrating on failure for instance. [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server [None req-f0f51e73-3699-448a-bf92-ae88bb9cc94a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Exception during message handling: AttributeError: 'NoneType' object has no attribute 'key' [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server raise self.value [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server raise self.value [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server raise self.value [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6071, in resize_instance [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server raise self.value [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6068, in resize_instance [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server self._resize_instance(context, instance, image, migration, [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server disk_info = self.driver.migrate_disk_and_power_off( [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 263, in migrate_disk_and_power_off [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server return self._vmops.migrate_disk_and_power_off(context, instance, [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server disk_key = device.key [ 1192.117291] env[61728]: ERROR oslo_messaging.rpc.server AttributeError: 'NoneType' object has no attribute 'key' [ 1192.119468] env[61728]: ERROR oslo_messaging.rpc.server [ 1192.177309] env[61728]: DEBUG nova.compute.manager [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1192.526564] env[61728]: DEBUG oslo_vmware.api [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465091, 'name': ReconfigVM_Task, 'duration_secs': 0.121659} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.527062] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122235', 'volume_id': '7bbf2464-ae28-4860-abbf-d14f03be449a', 'name': 'volume-7bbf2464-ae28-4860-abbf-d14f03be449a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6218d586-0190-4a02-ae88-3f74329dc32c', 'attached_at': '', 'detached_at': '', 'volume_id': '7bbf2464-ae28-4860-abbf-d14f03be449a', 'serial': '7bbf2464-ae28-4860-abbf-d14f03be449a'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1192.527489] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1192.528677] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b11643b-381a-4e1c-8725-c700b8f9c083 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.539711] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1192.540080] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7b460dc3-2a43-4079-afdf-4c807c2b7c42 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.632594] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1192.632868] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1192.633165] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleting the datastore file [datastore1] 6218d586-0190-4a02-ae88-3f74329dc32c {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1192.633504] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c3a60bc6-a6e1-4ce5-8c0d-c3a827e31956 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.640598] env[61728]: DEBUG oslo_vmware.api [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1192.640598] env[61728]: value = "task-465093" [ 1192.640598] env[61728]: _type = "Task" [ 1192.640598] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.648932] env[61728]: DEBUG oslo_vmware.api [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465093, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.701342] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1192.701615] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1192.703250] env[61728]: INFO nova.compute.claims [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1193.055392] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "2db96467-5fb4-49f4-954b-dbdd21e46232" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.055672] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "2db96467-5fb4-49f4-954b-dbdd21e46232" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.055928] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "2db96467-5fb4-49f4-954b-dbdd21e46232-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.056144] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "2db96467-5fb4-49f4-954b-dbdd21e46232-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.056326] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "2db96467-5fb4-49f4-954b-dbdd21e46232-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.058735] env[61728]: INFO nova.compute.manager [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Terminating instance [ 1193.060549] env[61728]: DEBUG nova.compute.manager [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1193.060749] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1193.061639] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df84cd2b-afb6-4a7c-966b-f4c5aabf0fbe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.070064] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1193.070289] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee2854f7-064e-4dd6-9763-e21764251922 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.077285] env[61728]: DEBUG oslo_vmware.api [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1193.077285] env[61728]: value = "task-465094" [ 1193.077285] env[61728]: _type = "Task" [ 1193.077285] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.084932] env[61728]: DEBUG oslo_vmware.api [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465094, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.151064] env[61728]: DEBUG oslo_vmware.api [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465093, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.098233} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.151338] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1193.151529] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1193.151713] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1193.151946] env[61728]: INFO nova.compute.manager [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Took 2.78 seconds to destroy the instance on the hypervisor. [ 1193.152245] env[61728]: DEBUG oslo.service.loopingcall [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1193.152474] env[61728]: DEBUG nova.compute.manager [-] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1193.152591] env[61728]: DEBUG nova.network.neutron [-] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1193.579215] env[61728]: DEBUG nova.compute.manager [req-beda5bed-2252-432b-9209-9e8886505264 req-9c99f2ea-d48a-4d0f-aa2e-e162a1910d71 service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Received event network-vif-deleted-e30c4bdc-f4f7-4bf7-bf54-d980230430e5 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1193.579486] env[61728]: INFO nova.compute.manager [req-beda5bed-2252-432b-9209-9e8886505264 req-9c99f2ea-d48a-4d0f-aa2e-e162a1910d71 service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Neutron deleted interface e30c4bdc-f4f7-4bf7-bf54-d980230430e5; detaching it from the instance and deleting it from the info cache [ 1193.579673] env[61728]: DEBUG nova.network.neutron [req-beda5bed-2252-432b-9209-9e8886505264 req-9c99f2ea-d48a-4d0f-aa2e-e162a1910d71 service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.590677] env[61728]: DEBUG oslo_vmware.api [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465094, 'name': PowerOffVM_Task, 'duration_secs': 0.198787} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.591418] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1193.591600] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1193.591864] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0eb5082a-d8d8-4cf8-8209-c90f48d9fabd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.660534] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1193.660765] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1193.660953] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleting the datastore file [datastore1] 2db96467-5fb4-49f4-954b-dbdd21e46232 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1193.661231] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7256daa6-50ef-4d36-8115-5cbcc8f99d9f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.668123] env[61728]: DEBUG oslo_vmware.api [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1193.668123] env[61728]: value = "task-465096" [ 1193.668123] env[61728]: _type = "Task" [ 1193.668123] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.676986] env[61728]: DEBUG oslo_vmware.api [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465096, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.795512] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b18308-3932-4191-ba46-d14a73fdc6fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.806596] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7e35b5-2e80-4315-a1c0-b3043604f7d3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.836826] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48de08d4-d8ad-4986-9a9c-47bb500cff43 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.844518] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f36d5772-c05e-4a58-970a-a3eb830bdf0d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.857748] env[61728]: DEBUG nova.compute.provider_tree [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1194.054904] env[61728]: DEBUG nova.network.neutron [-] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1194.085068] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-93ffd2a0-4e89-4bcd-a257-b273c871129c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.098558] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a3d16a-9c45-4397-a916-3798a3c87bc1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.126518] env[61728]: DEBUG nova.compute.manager [req-beda5bed-2252-432b-9209-9e8886505264 req-9c99f2ea-d48a-4d0f-aa2e-e162a1910d71 service nova] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Detach interface failed, port_id=e30c4bdc-f4f7-4bf7-bf54-d980230430e5, reason: Instance 6218d586-0190-4a02-ae88-3f74329dc32c could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1194.178255] env[61728]: DEBUG oslo_vmware.api [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465096, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137976} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.178527] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1194.178717] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1194.178903] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1194.179103] env[61728]: INFO nova.compute.manager [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1194.179358] env[61728]: DEBUG oslo.service.loopingcall [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1194.179565] env[61728]: DEBUG nova.compute.manager [-] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1194.179657] env[61728]: DEBUG nova.network.neutron [-] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1194.360754] env[61728]: DEBUG nova.scheduler.client.report [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1194.557962] env[61728]: INFO nova.compute.manager [-] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Took 1.41 seconds to deallocate network for instance. [ 1194.867067] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.165s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.867607] env[61728]: DEBUG nova.compute.manager [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1194.879225] env[61728]: DEBUG nova.network.neutron [-] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.103623] env[61728]: INFO nova.compute.manager [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Took 0.55 seconds to detach 1 volumes for instance. [ 1195.105913] env[61728]: DEBUG nova.compute.manager [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Deleting volume: 7bbf2464-ae28-4860-abbf-d14f03be449a {{(pid=61728) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1195.372548] env[61728]: DEBUG nova.compute.utils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1195.374240] env[61728]: DEBUG nova.compute.manager [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1195.374429] env[61728]: DEBUG nova.network.neutron [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1195.381119] env[61728]: INFO nova.compute.manager [-] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Took 1.20 seconds to deallocate network for instance. [ 1195.424096] env[61728]: DEBUG nova.policy [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e17b05d01da49e68afe0b7da5db6fb9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '833a646ea21a48c4951ed8cedebaa98f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1195.605711] env[61728]: DEBUG nova.compute.manager [req-f4508050-2f41-4aab-85b7-293b387ed7cf req-cd264fea-d957-456f-b5e8-0e1e31aba982 service nova] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Received event network-vif-deleted-e6562f3d-358f-482b-b0a1-b6a09c232c71 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1195.645528] env[61728]: DEBUG oslo_concurrency.lockutils [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.645806] env[61728]: DEBUG oslo_concurrency.lockutils [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.646075] env[61728]: DEBUG nova.objects.instance [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lazy-loading 'resources' on Instance uuid 6218d586-0190-4a02-ae88-3f74329dc32c {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1195.703050] env[61728]: DEBUG nova.network.neutron [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Successfully created port: e4ebecc4-1e02-4afe-aa2f-25362f481dc0 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1195.877907] env[61728]: DEBUG nova.compute.manager [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1195.887696] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.226605] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f5b169-786f-4865-bdd5-a3fe2929dfa2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.234632] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ba5fcd-c78f-4b49-8912-951c4f29ffa2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.264902] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e2a6b3-b1f2-4754-a04e-4679fee98f55 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.272187] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cef69e6-6a18-476e-be08-37543b5ecf54 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.285051] env[61728]: DEBUG nova.compute.provider_tree [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1196.788771] env[61728]: DEBUG nova.scheduler.client.report [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1196.887773] env[61728]: DEBUG nova.compute.manager [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1196.915897] env[61728]: DEBUG nova.virt.hardware [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1196.916170] env[61728]: DEBUG nova.virt.hardware [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1196.916448] env[61728]: DEBUG nova.virt.hardware [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1196.916512] env[61728]: DEBUG nova.virt.hardware [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1196.916667] env[61728]: DEBUG nova.virt.hardware [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1196.917582] env[61728]: DEBUG nova.virt.hardware [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1196.917582] env[61728]: DEBUG nova.virt.hardware [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1196.917582] env[61728]: DEBUG nova.virt.hardware [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1196.917582] env[61728]: DEBUG nova.virt.hardware [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1196.917582] env[61728]: DEBUG nova.virt.hardware [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1196.917815] env[61728]: DEBUG nova.virt.hardware [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1196.918579] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fc81bb4-a481-4054-afe3-c4fa11c9c752 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.926846] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8fa202-dd85-474c-a5de-92bcaaef0437 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.141137] env[61728]: DEBUG nova.network.neutron [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Successfully updated port: e4ebecc4-1e02-4afe-aa2f-25362f481dc0 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1197.295601] env[61728]: DEBUG oslo_concurrency.lockutils [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.648s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.296113] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.408s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.296362] env[61728]: DEBUG nova.objects.instance [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lazy-loading 'resources' on Instance uuid 2db96467-5fb4-49f4-954b-dbdd21e46232 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1197.317666] env[61728]: INFO nova.scheduler.client.report [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleted allocations for instance 6218d586-0190-4a02-ae88-3f74329dc32c [ 1197.630819] env[61728]: DEBUG nova.compute.manager [req-65714064-f177-4b49-af08-a6bdc9ad3b57 req-19a22b7d-2edb-4937-9214-99fcd2de301e service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Received event network-vif-plugged-e4ebecc4-1e02-4afe-aa2f-25362f481dc0 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1197.631076] env[61728]: DEBUG oslo_concurrency.lockutils [req-65714064-f177-4b49-af08-a6bdc9ad3b57 req-19a22b7d-2edb-4937-9214-99fcd2de301e service nova] Acquiring lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.631300] env[61728]: DEBUG oslo_concurrency.lockutils [req-65714064-f177-4b49-af08-a6bdc9ad3b57 req-19a22b7d-2edb-4937-9214-99fcd2de301e service nova] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.631475] env[61728]: DEBUG oslo_concurrency.lockutils [req-65714064-f177-4b49-af08-a6bdc9ad3b57 req-19a22b7d-2edb-4937-9214-99fcd2de301e service nova] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.631643] env[61728]: DEBUG nova.compute.manager [req-65714064-f177-4b49-af08-a6bdc9ad3b57 req-19a22b7d-2edb-4937-9214-99fcd2de301e service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] No waiting events found dispatching network-vif-plugged-e4ebecc4-1e02-4afe-aa2f-25362f481dc0 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1197.631832] env[61728]: WARNING nova.compute.manager [req-65714064-f177-4b49-af08-a6bdc9ad3b57 req-19a22b7d-2edb-4937-9214-99fcd2de301e service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Received unexpected event network-vif-plugged-e4ebecc4-1e02-4afe-aa2f-25362f481dc0 for instance with vm_state building and task_state spawning. [ 1197.632210] env[61728]: DEBUG nova.compute.manager [req-65714064-f177-4b49-af08-a6bdc9ad3b57 req-19a22b7d-2edb-4937-9214-99fcd2de301e service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Received event network-changed-e4ebecc4-1e02-4afe-aa2f-25362f481dc0 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1197.632405] env[61728]: DEBUG nova.compute.manager [req-65714064-f177-4b49-af08-a6bdc9ad3b57 req-19a22b7d-2edb-4937-9214-99fcd2de301e service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Refreshing instance network info cache due to event network-changed-e4ebecc4-1e02-4afe-aa2f-25362f481dc0. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1197.632598] env[61728]: DEBUG oslo_concurrency.lockutils [req-65714064-f177-4b49-af08-a6bdc9ad3b57 req-19a22b7d-2edb-4937-9214-99fcd2de301e service nova] Acquiring lock "refresh_cache-e5943edd-bdc3-4285-b2e0-7adac9acaf7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1197.632744] env[61728]: DEBUG oslo_concurrency.lockutils [req-65714064-f177-4b49-af08-a6bdc9ad3b57 req-19a22b7d-2edb-4937-9214-99fcd2de301e service nova] Acquired lock "refresh_cache-e5943edd-bdc3-4285-b2e0-7adac9acaf7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1197.632956] env[61728]: DEBUG nova.network.neutron [req-65714064-f177-4b49-af08-a6bdc9ad3b57 req-19a22b7d-2edb-4937-9214-99fcd2de301e service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Refreshing network info cache for port e4ebecc4-1e02-4afe-aa2f-25362f481dc0 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1197.644693] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "refresh_cache-e5943edd-bdc3-4285-b2e0-7adac9acaf7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1197.812471] env[61728]: DEBUG nova.compute.manager [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Stashing vm_state: active {{(pid=61728) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1197.824224] env[61728]: DEBUG oslo_concurrency.lockutils [None req-77731586-74ca-4de6-a16a-eea24f3732b0 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "6218d586-0190-4a02-ae88-3f74329dc32c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.461s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.825123] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "6218d586-0190-4a02-ae88-3f74329dc32c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.080s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.825323] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "6218d586-0190-4a02-ae88-3f74329dc32c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.825533] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "6218d586-0190-4a02-ae88-3f74329dc32c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.825694] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "6218d586-0190-4a02-ae88-3f74329dc32c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.827288] env[61728]: INFO nova.compute.manager [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Terminating instance [ 1197.831279] env[61728]: DEBUG nova.compute.manager [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1197.831997] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f3b1fe06-66c4-4e11-a8b5-8339e5135b2a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.841981] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e7e4e79-46d9-42e9-8bc7-1d9becbe95b3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.872688] env[61728]: WARNING nova.virt.vmwareapi.driver [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 6218d586-0190-4a02-ae88-3f74329dc32c could not be found. [ 1197.872921] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1197.873245] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04fbf70f-58c9-4cfe-bc73-2f2d0e02b1da {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.883906] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4243282-830c-428b-a61e-19ffbfc91bb5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.904875] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7e497c-9857-42cf-a63b-ba62a62746b3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.912221] env[61728]: WARNING nova.virt.vmwareapi.vmops [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6218d586-0190-4a02-ae88-3f74329dc32c could not be found. [ 1197.912453] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1197.912644] env[61728]: INFO nova.compute.manager [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Took 0.08 seconds to destroy the instance on the hypervisor. [ 1197.912885] env[61728]: DEBUG oslo.service.loopingcall [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1197.913447] env[61728]: DEBUG nova.compute.manager [-] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1197.913568] env[61728]: DEBUG nova.network.neutron [-] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1197.917418] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392b2be1-1ae4-4b0d-a229-93934f5d97d7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.949733] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a1978d-71db-4a10-bb09-98c9b997ccb6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.957807] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0b3496-17bd-40b8-a13f-4a40b55fc063 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.971899] env[61728]: DEBUG nova.compute.provider_tree [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1198.163772] env[61728]: DEBUG nova.network.neutron [req-65714064-f177-4b49-af08-a6bdc9ad3b57 req-19a22b7d-2edb-4937-9214-99fcd2de301e service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1198.238112] env[61728]: DEBUG nova.network.neutron [req-65714064-f177-4b49-af08-a6bdc9ad3b57 req-19a22b7d-2edb-4937-9214-99fcd2de301e service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.335283] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.453569] env[61728]: DEBUG nova.network.neutron [-] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.474950] env[61728]: DEBUG nova.scheduler.client.report [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1198.740436] env[61728]: DEBUG oslo_concurrency.lockutils [req-65714064-f177-4b49-af08-a6bdc9ad3b57 req-19a22b7d-2edb-4937-9214-99fcd2de301e service nova] Releasing lock "refresh_cache-e5943edd-bdc3-4285-b2e0-7adac9acaf7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1198.740771] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquired lock "refresh_cache-e5943edd-bdc3-4285-b2e0-7adac9acaf7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1198.740938] env[61728]: DEBUG nova.network.neutron [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1198.955941] env[61728]: INFO nova.compute.manager [-] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Took 1.04 seconds to deallocate network for instance. [ 1198.969225] env[61728]: WARNING nova.volume.cinder [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Attachment 0da3ce6e-9ad2-4601-920a-f60f464713fe does not exist. Ignoring.: cinderclient.exceptions.NotFound: Volume attachment could not be found with filter: attachment_id = 0da3ce6e-9ad2-4601-920a-f60f464713fe. (HTTP 404) (Request-ID: req-31b217f9-70a6-4f20-a801-02605ce6ec83) [ 1198.969528] env[61728]: INFO nova.compute.manager [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Took 0.01 seconds to detach 1 volumes for instance. [ 1198.972237] env[61728]: DEBUG nova.compute.manager [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Deleting volume: 7bbf2464-ae28-4860-abbf-d14f03be449a {{(pid=61728) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1198.979153] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.683s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.981393] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.646s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.987171] env[61728]: WARNING nova.compute.manager [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Failed to delete volume: 7bbf2464-ae28-4860-abbf-d14f03be449a due to Volume 7bbf2464-ae28-4860-abbf-d14f03be449a could not be found.: nova.exception.VolumeNotFound: Volume 7bbf2464-ae28-4860-abbf-d14f03be449a could not be found. [ 1199.002242] env[61728]: INFO nova.scheduler.client.report [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleted allocations for instance 2db96467-5fb4-49f4-954b-dbdd21e46232 [ 1199.270133] env[61728]: DEBUG nova.network.neutron [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1199.383538] env[61728]: DEBUG nova.network.neutron [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Updating instance_info_cache with network_info: [{"id": "e4ebecc4-1e02-4afe-aa2f-25362f481dc0", "address": "fa:16:3e:84:9c:d5", "network": {"id": "72d11e2a-907e-419e-b8a4-53aa39f2566a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1755247352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833a646ea21a48c4951ed8cedebaa98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3836fb52-19c6-4e10-a0ca-f0bca73dc887", "external-id": "nsx-vlan-transportzone-964", "segmentation_id": 964, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4ebecc4-1e", "ovs_interfaceid": "e4ebecc4-1e02-4afe-aa2f-25362f481dc0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.486804] env[61728]: INFO nova.compute.claims [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1199.490749] env[61728]: INFO nova.compute.manager [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Instance disappeared during terminate [ 1199.490954] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a3a521c3-567a-46f6-b679-2ea6ad7515e8 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "6218d586-0190-4a02-ae88-3f74329dc32c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 1.666s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.509018] env[61728]: DEBUG oslo_concurrency.lockutils [None req-cb62a2d1-5f34-474e-8d2f-8be49c865434 tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "2db96467-5fb4-49f4-954b-dbdd21e46232" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.453s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.885798] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Releasing lock "refresh_cache-e5943edd-bdc3-4285-b2e0-7adac9acaf7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1199.886168] env[61728]: DEBUG nova.compute.manager [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Instance network_info: |[{"id": "e4ebecc4-1e02-4afe-aa2f-25362f481dc0", "address": "fa:16:3e:84:9c:d5", "network": {"id": "72d11e2a-907e-419e-b8a4-53aa39f2566a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1755247352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833a646ea21a48c4951ed8cedebaa98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3836fb52-19c6-4e10-a0ca-f0bca73dc887", "external-id": "nsx-vlan-transportzone-964", "segmentation_id": 964, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4ebecc4-1e", "ovs_interfaceid": "e4ebecc4-1e02-4afe-aa2f-25362f481dc0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1199.886630] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:9c:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3836fb52-19c6-4e10-a0ca-f0bca73dc887', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e4ebecc4-1e02-4afe-aa2f-25362f481dc0', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1199.894189] env[61728]: DEBUG oslo.service.loopingcall [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1199.894401] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1199.894638] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eb071024-2fb0-4f4d-b202-f0ea15549068 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.915764] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1199.915764] env[61728]: value = "task-465098" [ 1199.915764] env[61728]: _type = "Task" [ 1199.915764] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.923469] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465098, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.992837] env[61728]: INFO nova.compute.resource_tracker [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating resource usage from migration 1c17fa7b-4e03-4cd8-a0d5-17fc0a716487 [ 1200.068755] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b20c6b9-f35e-42e8-a727-a5a0025b639f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.076349] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2288ee1a-b9ac-4b70-8b37-6b0c63495b41 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.106547] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ad6d5d-15e6-4073-9d16-4990d707a183 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.114435] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38972073-16d6-4be6-b746-2ebeef6bebfd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.128240] env[61728]: DEBUG nova.compute.provider_tree [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1200.426037] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465098, 'name': CreateVM_Task, 'duration_secs': 0.292481} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.426159] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1200.426839] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1200.427025] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1200.427366] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1200.427616] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7908b629-e51f-4256-98eb-6585927bfeb6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.432090] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1200.432090] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525f1b54-e1c6-6092-5b2f-78851f404488" [ 1200.432090] env[61728]: _type = "Task" [ 1200.432090] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.439370] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525f1b54-e1c6-6092-5b2f-78851f404488, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.632196] env[61728]: DEBUG nova.scheduler.client.report [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1200.942442] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]525f1b54-e1c6-6092-5b2f-78851f404488, 'name': SearchDatastore_Task, 'duration_secs': 0.011996} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.942705] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1200.942964] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1200.943227] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1200.943380] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1200.943571] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1200.943825] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-010677a3-8d96-45c8-9dee-65af6440d3ec {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.951869] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1200.952055] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1200.952714] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4014f17d-c896-417a-b0db-2b4e7f686dd4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.958987] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1200.958987] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52763343-8138-67ff-a648-5bdb14cd0bb1" [ 1200.958987] env[61728]: _type = "Task" [ 1200.958987] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.966112] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52763343-8138-67ff-a648-5bdb14cd0bb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.138361] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.157s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1201.138599] env[61728]: INFO nova.compute.manager [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Migrating [ 1201.470747] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52763343-8138-67ff-a648-5bdb14cd0bb1, 'name': SearchDatastore_Task, 'duration_secs': 0.008702} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.471679] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78183fa6-73c6-41d3-a98e-3b302b17eb91 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.477331] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1201.477331] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52793bab-5694-7277-2c31-7b32dd8b8d1e" [ 1201.477331] env[61728]: _type = "Task" [ 1201.477331] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.484574] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52793bab-5694-7277-2c31-7b32dd8b8d1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.653535] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1201.653776] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1201.654022] env[61728]: DEBUG nova.network.neutron [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1201.987431] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52793bab-5694-7277-2c31-7b32dd8b8d1e, 'name': SearchDatastore_Task, 'duration_secs': 0.009333} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.987642] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1201.987918] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] e5943edd-bdc3-4285-b2e0-7adac9acaf7f/e5943edd-bdc3-4285-b2e0-7adac9acaf7f.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1201.988196] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6844b68b-7962-44d6-afae-88a31076c3c0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.994540] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1201.994540] env[61728]: value = "task-465100" [ 1201.994540] env[61728]: _type = "Task" [ 1201.994540] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.002104] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465100, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.384352] env[61728]: DEBUG nova.network.neutron [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance_info_cache with network_info: [{"id": "8787e957-2e8a-4cce-b359-84b4b077d388", "address": "fa:16:3e:d6:47:92", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8787e957-2e", "ovs_interfaceid": "8787e957-2e8a-4cce-b359-84b4b077d388", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1202.504365] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465100, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.440509} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.504603] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] e5943edd-bdc3-4285-b2e0-7adac9acaf7f/e5943edd-bdc3-4285-b2e0-7adac9acaf7f.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1202.504828] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1202.505100] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a72118b0-3096-416d-a7fc-f5574bfe8c57 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.511151] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1202.511151] env[61728]: value = "task-465101" [ 1202.511151] env[61728]: _type = "Task" [ 1202.511151] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.518430] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465101, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.887290] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1203.020729] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465101, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073054} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.021065] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1203.021852] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef091c9-670d-4e9a-98ae-6ec4616cb4c4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.045607] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Reconfiguring VM instance instance-0000007b to attach disk [datastore1] e5943edd-bdc3-4285-b2e0-7adac9acaf7f/e5943edd-bdc3-4285-b2e0-7adac9acaf7f.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1203.045895] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b97038f-1b9b-47f2-bc50-b364efee7624 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.065378] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1203.065378] env[61728]: value = "task-465102" [ 1203.065378] env[61728]: _type = "Task" [ 1203.065378] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.072880] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465102, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.575438] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465102, 'name': ReconfigVM_Task, 'duration_secs': 0.30976} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.575764] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Reconfigured VM instance instance-0000007b to attach disk [datastore1] e5943edd-bdc3-4285-b2e0-7adac9acaf7f/e5943edd-bdc3-4285-b2e0-7adac9acaf7f.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1203.576394] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d0220bf1-baeb-409c-98f2-26a9120655ea {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.583288] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1203.583288] env[61728]: value = "task-465103" [ 1203.583288] env[61728]: _type = "Task" [ 1203.583288] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.591163] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465103, 'name': Rename_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.785561] env[61728]: DEBUG oslo_concurrency.lockutils [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.785836] env[61728]: DEBUG oslo_concurrency.lockutils [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.786080] env[61728]: DEBUG oslo_concurrency.lockutils [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.786279] env[61728]: DEBUG oslo_concurrency.lockutils [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.786457] env[61728]: DEBUG oslo_concurrency.lockutils [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.788826] env[61728]: INFO nova.compute.manager [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Terminating instance [ 1203.790695] env[61728]: DEBUG nova.compute.manager [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1203.790899] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1203.791798] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec1d54c-09d2-4465-90ee-55aca3294fd6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.799009] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1203.799249] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5c66253e-2c39-46cf-ae41-c13c76b9898b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.805644] env[61728]: DEBUG oslo_vmware.api [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1203.805644] env[61728]: value = "task-465104" [ 1203.805644] env[61728]: _type = "Task" [ 1203.805644] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.814517] env[61728]: DEBUG oslo_vmware.api [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465104, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.096263] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465103, 'name': Rename_Task, 'duration_secs': 0.144682} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.096477] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1204.096723] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55c948b9-f837-4b28-9ccd-5d93a4edc83d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.102850] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1204.102850] env[61728]: value = "task-465105" [ 1204.102850] env[61728]: _type = "Task" [ 1204.102850] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.110044] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465105, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.316180] env[61728]: DEBUG oslo_vmware.api [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465104, 'name': PowerOffVM_Task, 'duration_secs': 0.204141} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.316446] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1204.316621] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1204.316868] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-71b5d177-be7a-47f3-95ba-b120be02d8d4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.405748] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46103869-e2ad-485b-b97c-546bfd4de13d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.424952] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance 'fb1195b2-1ab5-47d7-be86-4b61e47b4e29' progress to 0 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1204.612530] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465105, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.930729] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1204.931080] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9adf9bc3-4763-4d25-9b19-82fb9bc99d5c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.938724] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1204.938724] env[61728]: value = "task-465107" [ 1204.938724] env[61728]: _type = "Task" [ 1204.938724] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.946453] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465107, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.114258] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465105, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.448725] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465107, 'name': PowerOffVM_Task, 'duration_secs': 0.226929} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.449046] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1205.449245] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance 'fb1195b2-1ab5-47d7-be86-4b61e47b4e29' progress to 17 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1205.615713] env[61728]: DEBUG oslo_vmware.api [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465105, 'name': PowerOnVM_Task, 'duration_secs': 1.152674} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.616088] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1205.616247] env[61728]: INFO nova.compute.manager [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Took 8.73 seconds to spawn the instance on the hypervisor. [ 1205.616429] env[61728]: DEBUG nova.compute.manager [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1205.617645] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e20b06-cfab-4f2f-8b97-3aca316944a1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.620782] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1205.620977] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1205.621182] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleting the datastore file [datastore1] b1fb2e3f-234a-44c7-983e-d4441b4e3cef {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1205.621890] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8b5820a8-9eff-4e63-bfc4-0b48c32bb2c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.629214] env[61728]: DEBUG oslo_vmware.api [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for the task: (returnval){ [ 1205.629214] env[61728]: value = "task-465108" [ 1205.629214] env[61728]: _type = "Task" [ 1205.629214] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.636634] env[61728]: DEBUG oslo_vmware.api [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465108, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.955589] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1205.955833] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1205.956012] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1205.956205] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1205.956359] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1205.956512] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1205.956725] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1205.956892] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1205.957084] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1205.957260] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1205.957442] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1205.962415] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e713a98-33b8-4b8e-af22-fea4eb05695d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.978120] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1205.978120] env[61728]: value = "task-465109" [ 1205.978120] env[61728]: _type = "Task" [ 1205.978120] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.985957] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465109, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.081717] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "532507fa-996d-45c7-bf70-f09c93be79ed" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1206.082066] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "532507fa-996d-45c7-bf70-f09c93be79ed" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1206.140412] env[61728]: INFO nova.compute.manager [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Took 13.46 seconds to build instance. [ 1206.144440] env[61728]: DEBUG oslo_vmware.api [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Task: {'id': task-465108, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155286} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.144786] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1206.145062] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1206.145368] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1206.145550] env[61728]: INFO nova.compute.manager [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Took 2.35 seconds to destroy the instance on the hypervisor. [ 1206.145820] env[61728]: DEBUG oslo.service.loopingcall [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1206.146115] env[61728]: DEBUG nova.compute.manager [-] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1206.146246] env[61728]: DEBUG nova.network.neutron [-] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1206.412292] env[61728]: DEBUG nova.compute.manager [req-26fac5b6-8b41-4b68-9254-07443518b1a9 req-3b5fea9d-d119-42b6-9251-14df8dfe6faa service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Received event network-vif-deleted-51b6c967-37bc-4156-a23f-91040bb1c4a2 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1206.412499] env[61728]: INFO nova.compute.manager [req-26fac5b6-8b41-4b68-9254-07443518b1a9 req-3b5fea9d-d119-42b6-9251-14df8dfe6faa service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Neutron deleted interface 51b6c967-37bc-4156-a23f-91040bb1c4a2; detaching it from the instance and deleting it from the info cache [ 1206.412675] env[61728]: DEBUG nova.network.neutron [req-26fac5b6-8b41-4b68-9254-07443518b1a9 req-3b5fea9d-d119-42b6-9251-14df8dfe6faa service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1206.489149] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465109, 'name': ReconfigVM_Task, 'duration_secs': 0.218974} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.489470] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance 'fb1195b2-1ab5-47d7-be86-4b61e47b4e29' progress to 33 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1206.585550] env[61728]: INFO nova.compute.manager [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Detaching volume ad2189db-dfcc-42fb-ab2a-15897b9ab1c8 [ 1206.619098] env[61728]: INFO nova.virt.block_device [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Attempting to driver detach volume ad2189db-dfcc-42fb-ab2a-15897b9ab1c8 from mountpoint /dev/sdb [ 1206.619379] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1206.619540] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122234', 'volume_id': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'name': 'volume-ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '532507fa-996d-45c7-bf70-f09c93be79ed', 'attached_at': '', 'detached_at': '', 'volume_id': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'serial': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1206.620520] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5990e955-1409-4a6e-ba4a-2b3c988b8b95 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.642414] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3c6643-d512-4e9e-a774-0b68e6fe7173 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.645730] env[61728]: DEBUG oslo_concurrency.lockutils [None req-ac0fdff6-35a9-4613-bffe-1fd0737ee199 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.971s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1206.650353] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89918ef-1e70-4682-aa17-88e390bc607a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.670738] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94f6f08-854d-4a32-9c91-d6494d8531cd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.685397] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] The volume has not been displaced from its original location: [datastore1] volume-ad2189db-dfcc-42fb-ab2a-15897b9ab1c8/volume-ad2189db-dfcc-42fb-ab2a-15897b9ab1c8.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1206.690839] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Reconfiguring VM instance instance-0000006f to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1206.691285] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75177354-3537-456c-907d-d7af9dae62b6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.710097] env[61728]: DEBUG oslo_vmware.api [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1206.710097] env[61728]: value = "task-465110" [ 1206.710097] env[61728]: _type = "Task" [ 1206.710097] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.718059] env[61728]: DEBUG oslo_vmware.api [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465110, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.888648] env[61728]: DEBUG nova.network.neutron [-] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1206.915539] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b29ac8f-70a2-4d96-b624-c430ad299204 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.925800] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b077a35-5116-4807-b108-0a7d2650dfd1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.940887] env[61728]: DEBUG nova.compute.manager [req-88063a5f-e27c-455b-8ce2-445282e9eb2f req-4c199838-c192-4f7c-99f7-71f01cdc5e63 service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Received event network-changed-e4ebecc4-1e02-4afe-aa2f-25362f481dc0 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1206.941146] env[61728]: DEBUG nova.compute.manager [req-88063a5f-e27c-455b-8ce2-445282e9eb2f req-4c199838-c192-4f7c-99f7-71f01cdc5e63 service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Refreshing instance network info cache due to event network-changed-e4ebecc4-1e02-4afe-aa2f-25362f481dc0. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1206.941365] env[61728]: DEBUG oslo_concurrency.lockutils [req-88063a5f-e27c-455b-8ce2-445282e9eb2f req-4c199838-c192-4f7c-99f7-71f01cdc5e63 service nova] Acquiring lock "refresh_cache-e5943edd-bdc3-4285-b2e0-7adac9acaf7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1206.941542] env[61728]: DEBUG oslo_concurrency.lockutils [req-88063a5f-e27c-455b-8ce2-445282e9eb2f req-4c199838-c192-4f7c-99f7-71f01cdc5e63 service nova] Acquired lock "refresh_cache-e5943edd-bdc3-4285-b2e0-7adac9acaf7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1206.941675] env[61728]: DEBUG nova.network.neutron [req-88063a5f-e27c-455b-8ce2-445282e9eb2f req-4c199838-c192-4f7c-99f7-71f01cdc5e63 service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Refreshing network info cache for port e4ebecc4-1e02-4afe-aa2f-25362f481dc0 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1206.958327] env[61728]: DEBUG nova.compute.manager [req-26fac5b6-8b41-4b68-9254-07443518b1a9 req-3b5fea9d-d119-42b6-9251-14df8dfe6faa service nova] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Detach interface failed, port_id=51b6c967-37bc-4156-a23f-91040bb1c4a2, reason: Instance b1fb2e3f-234a-44c7-983e-d4441b4e3cef could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1206.995272] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1206.995481] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1206.995644] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1206.995832] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1206.995988] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1206.996155] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1206.996361] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1206.996526] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1206.996697] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1206.996863] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1206.997052] env[61728]: DEBUG nova.virt.hardware [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1207.002338] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Reconfiguring VM instance instance-00000078 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1207.002640] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d52e99c1-2cb4-48cf-ac28-f52cb6e6a972 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.022436] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1207.022436] env[61728]: value = "task-465111" [ 1207.022436] env[61728]: _type = "Task" [ 1207.022436] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.030892] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465111, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.218978] env[61728]: DEBUG oslo_vmware.api [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465110, 'name': ReconfigVM_Task, 'duration_secs': 0.27414} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.219399] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Reconfigured VM instance instance-0000006f to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1207.224067] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dec362b7-0021-4776-a003-45ed07a1f6d4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.239122] env[61728]: DEBUG oslo_vmware.api [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1207.239122] env[61728]: value = "task-465112" [ 1207.239122] env[61728]: _type = "Task" [ 1207.239122] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.246967] env[61728]: DEBUG oslo_vmware.api [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465112, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.391625] env[61728]: INFO nova.compute.manager [-] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Took 1.25 seconds to deallocate network for instance. [ 1207.532337] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465111, 'name': ReconfigVM_Task, 'duration_secs': 0.168853} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.534617] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Reconfigured VM instance instance-00000078 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1207.535407] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a80472-2c2a-47f5-9ffc-763dfb5c8eeb {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.557321] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Reconfiguring VM instance instance-00000078 to attach disk [datastore1] fb1195b2-1ab5-47d7-be86-4b61e47b4e29/fb1195b2-1ab5-47d7-be86-4b61e47b4e29.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1207.559703] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b89b9f7-a21a-4bcd-a3cc-757326d7217d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.578091] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1207.578091] env[61728]: value = "task-465113" [ 1207.578091] env[61728]: _type = "Task" [ 1207.578091] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.585914] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465113, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.663669] env[61728]: DEBUG nova.network.neutron [req-88063a5f-e27c-455b-8ce2-445282e9eb2f req-4c199838-c192-4f7c-99f7-71f01cdc5e63 service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Updated VIF entry in instance network info cache for port e4ebecc4-1e02-4afe-aa2f-25362f481dc0. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1207.664081] env[61728]: DEBUG nova.network.neutron [req-88063a5f-e27c-455b-8ce2-445282e9eb2f req-4c199838-c192-4f7c-99f7-71f01cdc5e63 service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Updating instance_info_cache with network_info: [{"id": "e4ebecc4-1e02-4afe-aa2f-25362f481dc0", "address": "fa:16:3e:84:9c:d5", "network": {"id": "72d11e2a-907e-419e-b8a4-53aa39f2566a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1755247352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833a646ea21a48c4951ed8cedebaa98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3836fb52-19c6-4e10-a0ca-f0bca73dc887", "external-id": "nsx-vlan-transportzone-964", "segmentation_id": 964, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4ebecc4-1e", "ovs_interfaceid": "e4ebecc4-1e02-4afe-aa2f-25362f481dc0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1207.749818] env[61728]: DEBUG oslo_vmware.api [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465112, 'name': ReconfigVM_Task, 'duration_secs': 0.150773} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.750185] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122234', 'volume_id': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'name': 'volume-ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '532507fa-996d-45c7-bf70-f09c93be79ed', 'attached_at': '', 'detached_at': '', 'volume_id': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8', 'serial': 'ad2189db-dfcc-42fb-ab2a-15897b9ab1c8'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1207.898434] env[61728]: DEBUG oslo_concurrency.lockutils [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.899119] env[61728]: DEBUG oslo_concurrency.lockutils [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.899119] env[61728]: DEBUG nova.objects.instance [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lazy-loading 'resources' on Instance uuid b1fb2e3f-234a-44c7-983e-d4441b4e3cef {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.088390] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465113, 'name': ReconfigVM_Task, 'duration_secs': 0.318962} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.088683] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Reconfigured VM instance instance-00000078 to attach disk [datastore1] fb1195b2-1ab5-47d7-be86-4b61e47b4e29/fb1195b2-1ab5-47d7-be86-4b61e47b4e29.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1208.088968] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance 'fb1195b2-1ab5-47d7-be86-4b61e47b4e29' progress to 50 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1208.166549] env[61728]: DEBUG oslo_concurrency.lockutils [req-88063a5f-e27c-455b-8ce2-445282e9eb2f req-4c199838-c192-4f7c-99f7-71f01cdc5e63 service nova] Releasing lock "refresh_cache-e5943edd-bdc3-4285-b2e0-7adac9acaf7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1208.296172] env[61728]: DEBUG nova.objects.instance [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lazy-loading 'flavor' on Instance uuid 532507fa-996d-45c7-bf70-f09c93be79ed {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.473796] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31faac94-9c58-4fa2-909a-1217914e14bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.481586] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd89ab1-993e-4980-b50d-c7aedc83ef98 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.511357] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c669907a-b22c-4f91-bbf5-bb72c61d70ad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.518099] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9fb7289-d6dc-4af8-91f2-a3d088df7a7b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.531758] env[61728]: DEBUG nova.compute.provider_tree [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1208.595973] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8c4569-023c-4bcd-ae19-1c389bb007b3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.614786] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ef0518-7916-4080-9b4a-3833a0f096f1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.633175] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance 'fb1195b2-1ab5-47d7-be86-4b61e47b4e29' progress to 67 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1209.034688] env[61728]: DEBUG nova.scheduler.client.report [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1209.172469] env[61728]: DEBUG nova.network.neutron [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Port 8787e957-2e8a-4cce-b359-84b4b077d388 binding to destination host cpu-1 is already ACTIVE {{(pid=61728) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1209.303608] env[61728]: DEBUG oslo_concurrency.lockutils [None req-d7966408-3a09-4c30-960d-757acaeea214 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "532507fa-996d-45c7-bf70-f09c93be79ed" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.221s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.539438] env[61728]: DEBUG oslo_concurrency.lockutils [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.640s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.560550] env[61728]: INFO nova.scheduler.client.report [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Deleted allocations for instance b1fb2e3f-234a-44c7-983e-d4441b4e3cef [ 1210.068863] env[61728]: DEBUG oslo_concurrency.lockutils [None req-078ea37f-bd05-4902-ab92-140dccd13a8a tempest-ServerActionsTestOtherA-1690531141 tempest-ServerActionsTestOtherA-1690531141-project-member] Lock "b1fb2e3f-234a-44c7-983e-d4441b4e3cef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.283s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.200228] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.200228] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.200228] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.372645] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "532507fa-996d-45c7-bf70-f09c93be79ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.372895] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "532507fa-996d-45c7-bf70-f09c93be79ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.373158] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "532507fa-996d-45c7-bf70-f09c93be79ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.373356] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "532507fa-996d-45c7-bf70-f09c93be79ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.373532] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "532507fa-996d-45c7-bf70-f09c93be79ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.375656] env[61728]: INFO nova.compute.manager [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Terminating instance [ 1210.377517] env[61728]: DEBUG nova.compute.manager [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1210.377714] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1210.378605] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea5c387-03d5-48b8-8001-5fae8123703a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.387685] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1210.387916] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ab9b6fa5-e385-4a3a-9aba-3959e4d85a99 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.394257] env[61728]: DEBUG oslo_vmware.api [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1210.394257] env[61728]: value = "task-465114" [ 1210.394257] env[61728]: _type = "Task" [ 1210.394257] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.402019] env[61728]: DEBUG oslo_vmware.api [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465114, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.904274] env[61728]: DEBUG oslo_vmware.api [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465114, 'name': PowerOffVM_Task, 'duration_secs': 0.18911} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.905126] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1210.905348] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1210.905600] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb75c30d-686d-452c-82df-5d57bbcd7715 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.242369] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1211.242603] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1211.242757] env[61728]: DEBUG nova.network.neutron [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1211.701591] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1211.701917] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1211.702220] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleting the datastore file [datastore1] 532507fa-996d-45c7-bf70-f09c93be79ed {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1211.702478] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9bde6778-4f40-4bd6-8092-fa65f0e852f3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.708814] env[61728]: DEBUG oslo_vmware.api [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1211.708814] env[61728]: value = "task-465116" [ 1211.708814] env[61728]: _type = "Task" [ 1211.708814] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.716353] env[61728]: DEBUG oslo_vmware.api [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465116, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.960480] env[61728]: DEBUG nova.network.neutron [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance_info_cache with network_info: [{"id": "8787e957-2e8a-4cce-b359-84b4b077d388", "address": "fa:16:3e:d6:47:92", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8787e957-2e", "ovs_interfaceid": "8787e957-2e8a-4cce-b359-84b4b077d388", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.219530] env[61728]: DEBUG oslo_vmware.api [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465116, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17883} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.219949] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1212.220239] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1212.220477] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1212.220682] env[61728]: INFO nova.compute.manager [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Took 1.84 seconds to destroy the instance on the hypervisor. [ 1212.220949] env[61728]: DEBUG oslo.service.loopingcall [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1212.221173] env[61728]: DEBUG nova.compute.manager [-] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1212.221269] env[61728]: DEBUG nova.network.neutron [-] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1212.463311] env[61728]: DEBUG oslo_concurrency.lockutils [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1212.639244] env[61728]: DEBUG nova.compute.manager [req-be6a1c94-6bb0-4946-bc89-823a41839d8e req-7d9bcb66-d951-4b77-bfbb-4ae092771288 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Received event network-vif-deleted-d04bc3a5-d378-4e1e-a145-ec523bcc442a {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1212.639463] env[61728]: INFO nova.compute.manager [req-be6a1c94-6bb0-4946-bc89-823a41839d8e req-7d9bcb66-d951-4b77-bfbb-4ae092771288 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Neutron deleted interface d04bc3a5-d378-4e1e-a145-ec523bcc442a; detaching it from the instance and deleting it from the info cache [ 1212.639641] env[61728]: DEBUG nova.network.neutron [req-be6a1c94-6bb0-4946-bc89-823a41839d8e req-7d9bcb66-d951-4b77-bfbb-4ae092771288 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.988850] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-927575d2-89e6-47c8-aed3-7692102297c4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.007904] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabce9e6-07e3-41da-b6e7-5fecd44ec275 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.014938] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance 'fb1195b2-1ab5-47d7-be86-4b61e47b4e29' progress to 83 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1213.118307] env[61728]: DEBUG nova.network.neutron [-] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1213.141947] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a1d86216-3ca7-42e5-9354-c39a29ef8bcd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.151929] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7da257-470a-449f-a76a-4f9c4f047866 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.176517] env[61728]: DEBUG nova.compute.manager [req-be6a1c94-6bb0-4946-bc89-823a41839d8e req-7d9bcb66-d951-4b77-bfbb-4ae092771288 service nova] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Detach interface failed, port_id=d04bc3a5-d378-4e1e-a145-ec523bcc442a, reason: Instance 532507fa-996d-45c7-bf70-f09c93be79ed could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1213.521589] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1213.521989] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c05c7092-061e-41c2-a13c-2a937eb35d1a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.530346] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1213.530346] env[61728]: value = "task-465117" [ 1213.530346] env[61728]: _type = "Task" [ 1213.530346] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.538824] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465117, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.620578] env[61728]: INFO nova.compute.manager [-] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Took 1.40 seconds to deallocate network for instance. [ 1214.041013] env[61728]: DEBUG oslo_vmware.api [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465117, 'name': PowerOnVM_Task, 'duration_secs': 0.474815} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.041303] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1214.041494] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-9792365e-d629-40a1-ac38-030e9d59c00b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance 'fb1195b2-1ab5-47d7-be86-4b61e47b4e29' progress to 100 {{(pid=61728) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1214.127081] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.127370] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.127599] env[61728]: DEBUG nova.objects.instance [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lazy-loading 'resources' on Instance uuid 532507fa-996d-45c7-bf70-f09c93be79ed {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1214.686662] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fe7bf9-270e-4cdd-b8ed-ac32a502c854 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.694770] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9dcb17-a6a4-4654-8ef7-33bfdd432fb5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.725247] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffac25c5-c692-4a4d-b4a7-4c74fe2fc024 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.732074] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fb68f9c-c784-4669-abad-9b034f3d5e79 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.744718] env[61728]: DEBUG nova.compute.provider_tree [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1215.247114] env[61728]: DEBUG nova.scheduler.client.report [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1215.753091] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.625s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.770955] env[61728]: INFO nova.scheduler.client.report [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleted allocations for instance 532507fa-996d-45c7-bf70-f09c93be79ed [ 1216.278269] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2f2f1dbb-f001-482b-9cae-46137cf1abe7 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "532507fa-996d-45c7-bf70-f09c93be79ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.905s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.465014] env[61728]: DEBUG nova.network.neutron [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Port 8787e957-2e8a-4cce-b359-84b4b077d388 binding to destination host cpu-1 is already ACTIVE {{(pid=61728) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1216.465324] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1216.465547] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1216.465683] env[61728]: DEBUG nova.network.neutron [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1217.182262] env[61728]: DEBUG nova.network.neutron [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance_info_cache with network_info: [{"id": "8787e957-2e8a-4cce-b359-84b4b077d388", "address": "fa:16:3e:d6:47:92", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8787e957-2e", "ovs_interfaceid": "8787e957-2e8a-4cce-b359-84b4b077d388", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1217.684991] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1218.188866] env[61728]: DEBUG nova.compute.manager [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61728) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1218.189169] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.189357] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.421077] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "f95e28ce-0792-488a-a730-0766272ab2d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.421270] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "f95e28ce-0792-488a-a730-0766272ab2d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.692476] env[61728]: DEBUG nova.objects.instance [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lazy-loading 'migration_context' on Instance uuid fb1195b2-1ab5-47d7-be86-4b61e47b4e29 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1218.923414] env[61728]: DEBUG nova.compute.manager [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1219.256100] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24d5625-3c08-4856-9efa-9b959c154924 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.264298] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462d3ff3-3e28-490b-87db-92984b85fefa {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.293335] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33dbc05f-dfdc-403e-b9f8-a35146904852 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.300816] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760b49ef-0744-4bb6-8edd-4b06c8ae2730 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.315140] env[61728]: DEBUG nova.compute.provider_tree [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1219.441139] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1219.818774] env[61728]: DEBUG nova.scheduler.client.report [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1220.830656] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.641s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1220.836439] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.395s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1220.837926] env[61728]: INFO nova.compute.claims [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1221.902437] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e2adc5-5734-47b3-9d79-d2a3ace728d3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.909970] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1fe7485-a280-4802-9d81-c299c9366e7f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.940433] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56b36e5-ea53-4d8f-b621-28ca692cfdfd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.947334] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d21a98-de9a-45ee-b9db-6cda8b26f830 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.959830] env[61728]: DEBUG nova.compute.provider_tree [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1222.370307] env[61728]: INFO nova.compute.manager [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Swapping old allocation on dict_keys(['e7ceb92f-072b-409e-b888-6fe0676b32f1']) held by migration 1c17fa7b-4e03-4cd8-a0d5-17fc0a716487 for instance [ 1222.390622] env[61728]: DEBUG nova.scheduler.client.report [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Overwriting current allocation {'allocations': {'e7ceb92f-072b-409e-b888-6fe0676b32f1': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 199}}, 'project_id': 'f7f61c87df8b485b9a6370032dd7af37', 'user_id': '67d7c6a6191045cf85f790acebaf811e', 'consumer_generation': 1} on consumer fb1195b2-1ab5-47d7-be86-4b61e47b4e29 {{(pid=61728) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1222.462906] env[61728]: DEBUG nova.scheduler.client.report [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1222.466955] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1222.467148] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.467325] env[61728]: DEBUG nova.network.neutron [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1222.969680] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.133s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1222.970265] env[61728]: DEBUG nova.compute.manager [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1223.346928] env[61728]: DEBUG nova.network.neutron [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance_info_cache with network_info: [{"id": "8787e957-2e8a-4cce-b359-84b4b077d388", "address": "fa:16:3e:d6:47:92", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8787e957-2e", "ovs_interfaceid": "8787e957-2e8a-4cce-b359-84b4b077d388", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1223.476584] env[61728]: DEBUG nova.compute.utils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1223.478157] env[61728]: DEBUG nova.compute.manager [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1223.478332] env[61728]: DEBUG nova.network.neutron [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1223.514645] env[61728]: DEBUG nova.policy [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '836bc29cb90f4b3fb2459041a65805c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '438de94d086e47b99cedc6e4a33471bf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1223.783126] env[61728]: DEBUG nova.network.neutron [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Successfully created port: a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1223.850086] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-fb1195b2-1ab5-47d7-be86-4b61e47b4e29" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1223.850695] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1223.850866] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72b85c54-65ab-4334-8166-a2192d9d15fc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.858374] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1223.858374] env[61728]: value = "task-465118" [ 1223.858374] env[61728]: _type = "Task" [ 1223.858374] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.866550] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465118, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.981509] env[61728]: DEBUG nova.compute.manager [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1224.368757] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465118, 'name': PowerOffVM_Task, 'duration_secs': 0.24085} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.369052] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1224.369731] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1224.369999] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1224.370186] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1224.370382] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1224.370535] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1224.370688] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1224.370972] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1224.371199] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1224.371406] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1224.371581] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1224.371759] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1224.376878] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2c1b187-01a5-41bc-928b-f898f01879d6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.394310] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1224.394310] env[61728]: value = "task-465119" [ 1224.394310] env[61728]: _type = "Task" [ 1224.394310] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.403136] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465119, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.905473] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465119, 'name': ReconfigVM_Task, 'duration_secs': 0.236278} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.906352] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a8356d-e63c-44ad-9942-55914710995f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.923971] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1224.924253] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1224.924419] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1224.924609] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1224.924762] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1224.924919] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1224.925145] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1224.925314] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1224.925488] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1224.925658] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1224.925837] env[61728]: DEBUG nova.virt.hardware [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1224.926677] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e79c3559-251b-4e07-ba26-762f405942d4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.932066] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1224.932066] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527636bc-0580-c1e6-eaba-53aef405ca3e" [ 1224.932066] env[61728]: _type = "Task" [ 1224.932066] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.939775] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527636bc-0580-c1e6-eaba-53aef405ca3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.992279] env[61728]: DEBUG nova.compute.manager [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1225.017455] env[61728]: DEBUG nova.virt.hardware [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1225.017702] env[61728]: DEBUG nova.virt.hardware [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1225.017869] env[61728]: DEBUG nova.virt.hardware [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1225.018073] env[61728]: DEBUG nova.virt.hardware [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1225.018234] env[61728]: DEBUG nova.virt.hardware [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1225.018388] env[61728]: DEBUG nova.virt.hardware [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1225.018600] env[61728]: DEBUG nova.virt.hardware [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1225.018763] env[61728]: DEBUG nova.virt.hardware [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1225.018962] env[61728]: DEBUG nova.virt.hardware [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1225.019112] env[61728]: DEBUG nova.virt.hardware [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1225.019294] env[61728]: DEBUG nova.virt.hardware [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1225.020217] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb91e30-63d0-484f-adae-cce1ecf93691 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.029366] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76fe643-5c82-4e3c-877c-a33eddb46d03 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.146361] env[61728]: DEBUG nova.compute.manager [req-1bf3bd21-f6b5-415c-9cae-efe45f27716f req-511389cc-1c6a-4193-92e4-222c8d4ec346 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Received event network-vif-plugged-a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1225.146361] env[61728]: DEBUG oslo_concurrency.lockutils [req-1bf3bd21-f6b5-415c-9cae-efe45f27716f req-511389cc-1c6a-4193-92e4-222c8d4ec346 service nova] Acquiring lock "f95e28ce-0792-488a-a730-0766272ab2d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1225.146361] env[61728]: DEBUG oslo_concurrency.lockutils [req-1bf3bd21-f6b5-415c-9cae-efe45f27716f req-511389cc-1c6a-4193-92e4-222c8d4ec346 service nova] Lock "f95e28ce-0792-488a-a730-0766272ab2d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1225.146361] env[61728]: DEBUG oslo_concurrency.lockutils [req-1bf3bd21-f6b5-415c-9cae-efe45f27716f req-511389cc-1c6a-4193-92e4-222c8d4ec346 service nova] Lock "f95e28ce-0792-488a-a730-0766272ab2d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1225.146361] env[61728]: DEBUG nova.compute.manager [req-1bf3bd21-f6b5-415c-9cae-efe45f27716f req-511389cc-1c6a-4193-92e4-222c8d4ec346 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] No waiting events found dispatching network-vif-plugged-a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1225.146361] env[61728]: WARNING nova.compute.manager [req-1bf3bd21-f6b5-415c-9cae-efe45f27716f req-511389cc-1c6a-4193-92e4-222c8d4ec346 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Received unexpected event network-vif-plugged-a3aff6d0-5474-4d96-9da8-d7b907af2138 for instance with vm_state building and task_state spawning. [ 1225.221351] env[61728]: DEBUG nova.network.neutron [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Successfully updated port: a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1225.441842] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]527636bc-0580-c1e6-eaba-53aef405ca3e, 'name': SearchDatastore_Task, 'duration_secs': 0.009944} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.447056] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Reconfiguring VM instance instance-00000078 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1225.447339] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6cf0155-604f-45c0-94c7-140e6634b3c5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.465324] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1225.465324] env[61728]: value = "task-465120" [ 1225.465324] env[61728]: _type = "Task" [ 1225.465324] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.473269] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465120, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.724323] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1225.724562] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1225.724626] env[61728]: DEBUG nova.network.neutron [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1225.975483] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465120, 'name': ReconfigVM_Task, 'duration_secs': 0.202854} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.975709] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Reconfigured VM instance instance-00000078 to detach disk 2000 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1225.976478] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf53cb6-e663-4725-9e93-3e44f8a0046e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.999523] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Reconfiguring VM instance instance-00000078 to attach disk [datastore1] fb1195b2-1ab5-47d7-be86-4b61e47b4e29/fb1195b2-1ab5-47d7-be86-4b61e47b4e29.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1225.999927] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5cb4d50-e1ee-4c4e-b00f-a660dd58efb3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.017440] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1226.017440] env[61728]: value = "task-465121" [ 1226.017440] env[61728]: _type = "Task" [ 1226.017440] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.025184] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465121, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.255073] env[61728]: DEBUG nova.network.neutron [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1226.373871] env[61728]: DEBUG nova.network.neutron [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updating instance_info_cache with network_info: [{"id": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "address": "fa:16:3e:30:bd:13", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3aff6d0-54", "ovs_interfaceid": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1226.526903] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465121, 'name': ReconfigVM_Task, 'duration_secs': 0.269204} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.527219] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Reconfigured VM instance instance-00000078 to attach disk [datastore1] fb1195b2-1ab5-47d7-be86-4b61e47b4e29/fb1195b2-1ab5-47d7-be86-4b61e47b4e29.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1226.528035] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae13dd58-98e9-4f60-ad30-7f75fe2dbc6d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.545617] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437916c1-f4d5-4d43-82b5-e600af1ac4ca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.562936] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb0821f8-c451-4208-b91c-ed145254da8c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.579385] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69b82af-0000-4633-970a-3e38e24b640c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.585725] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1226.585945] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55c4e26b-01f2-43a1-8873-a0d3a7d29cca {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.591571] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1226.591571] env[61728]: value = "task-465122" [ 1226.591571] env[61728]: _type = "Task" [ 1226.591571] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.598448] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465122, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.876968] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1226.877275] env[61728]: DEBUG nova.compute.manager [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Instance network_info: |[{"id": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "address": "fa:16:3e:30:bd:13", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3aff6d0-54", "ovs_interfaceid": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1226.877721] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:bd:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f856fca-9fb5-41ea-a057-ac4193bd323d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a3aff6d0-5474-4d96-9da8-d7b907af2138', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1226.885255] env[61728]: DEBUG oslo.service.loopingcall [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1226.885451] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1226.885680] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-96f3fed3-4508-47b7-ae08-73f36cd72f32 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.905337] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1226.905337] env[61728]: value = "task-465123" [ 1226.905337] env[61728]: _type = "Task" [ 1226.905337] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.912460] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465123, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.102431] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465122, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.169061] env[61728]: DEBUG nova.compute.manager [req-9b4c2cff-bea5-4203-a1b9-21e1597663fc req-60542226-a514-4218-ad5b-e1191845920a service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Received event network-changed-a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1227.169253] env[61728]: DEBUG nova.compute.manager [req-9b4c2cff-bea5-4203-a1b9-21e1597663fc req-60542226-a514-4218-ad5b-e1191845920a service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Refreshing instance network info cache due to event network-changed-a3aff6d0-5474-4d96-9da8-d7b907af2138. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1227.169455] env[61728]: DEBUG oslo_concurrency.lockutils [req-9b4c2cff-bea5-4203-a1b9-21e1597663fc req-60542226-a514-4218-ad5b-e1191845920a service nova] Acquiring lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1227.169603] env[61728]: DEBUG oslo_concurrency.lockutils [req-9b4c2cff-bea5-4203-a1b9-21e1597663fc req-60542226-a514-4218-ad5b-e1191845920a service nova] Acquired lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1227.169782] env[61728]: DEBUG nova.network.neutron [req-9b4c2cff-bea5-4203-a1b9-21e1597663fc req-60542226-a514-4218-ad5b-e1191845920a service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Refreshing network info cache for port a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1227.415393] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465123, 'name': CreateVM_Task} progress is 25%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.601219] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465122, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.845731] env[61728]: DEBUG nova.network.neutron [req-9b4c2cff-bea5-4203-a1b9-21e1597663fc req-60542226-a514-4218-ad5b-e1191845920a service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updated VIF entry in instance network info cache for port a3aff6d0-5474-4d96-9da8-d7b907af2138. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1227.846117] env[61728]: DEBUG nova.network.neutron [req-9b4c2cff-bea5-4203-a1b9-21e1597663fc req-60542226-a514-4218-ad5b-e1191845920a service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updating instance_info_cache with network_info: [{"id": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "address": "fa:16:3e:30:bd:13", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3aff6d0-54", "ovs_interfaceid": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1227.915989] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465123, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.102136] env[61728]: DEBUG oslo_vmware.api [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465122, 'name': PowerOnVM_Task, 'duration_secs': 1.328636} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.102426] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1228.349036] env[61728]: DEBUG oslo_concurrency.lockutils [req-9b4c2cff-bea5-4203-a1b9-21e1597663fc req-60542226-a514-4218-ad5b-e1191845920a service nova] Releasing lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1228.416855] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465123, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.917153] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465123, 'name': CreateVM_Task} progress is 99%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.114115] env[61728]: INFO nova.compute.manager [None req-b55abaec-160b-48a8-aba2-805880b3eac5 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance to original state: 'active' [ 1229.417737] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465123, 'name': CreateVM_Task, 'duration_secs': 2.333594} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.417989] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1229.418679] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1229.418819] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1229.419165] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1229.419415] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e89d8382-ae2e-4431-997d-86b56b810170 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.423800] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1229.423800] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5237948f-9da0-9941-0dfb-57aace2d4ada" [ 1229.423800] env[61728]: _type = "Task" [ 1229.423800] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.431093] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5237948f-9da0-9941-0dfb-57aace2d4ada, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.934427] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5237948f-9da0-9941-0dfb-57aace2d4ada, 'name': SearchDatastore_Task, 'duration_secs': 0.012626} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.934427] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1229.934502] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1229.934747] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1229.934915] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1229.935121] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1229.935382] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d6275e96-04ee-455d-a31e-4554f4d238e4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.942994] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1229.943191] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1229.943864] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71db1853-7d17-4a5f-b01b-cf3589245c7c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.948716] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1229.948716] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f77be7-6d2a-c991-033a-6e9503589fbf" [ 1229.948716] env[61728]: _type = "Task" [ 1229.948716] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.956101] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f77be7-6d2a-c991-033a-6e9503589fbf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.250943] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1230.251204] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.251423] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1230.251613] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.251800] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1230.254300] env[61728]: INFO nova.compute.manager [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Terminating instance [ 1230.255988] env[61728]: DEBUG nova.compute.manager [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1230.256170] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1230.256984] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9036cc12-dc3f-4077-8747-781922d2f310 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.264665] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1230.264893] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3c241988-7b3e-4cf3-83a2-171de32e712e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.271743] env[61728]: DEBUG oslo_vmware.api [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1230.271743] env[61728]: value = "task-465124" [ 1230.271743] env[61728]: _type = "Task" [ 1230.271743] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.278911] env[61728]: DEBUG oslo_vmware.api [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465124, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.458943] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52f77be7-6d2a-c991-033a-6e9503589fbf, 'name': SearchDatastore_Task, 'duration_secs': 0.008996} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.459881] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afe22150-8941-4146-b65f-a017a70ffb82 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.464772] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1230.464772] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c5381a-2228-ff56-312c-b1940becf742" [ 1230.464772] env[61728]: _type = "Task" [ 1230.464772] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.472022] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c5381a-2228-ff56-312c-b1940becf742, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.781268] env[61728]: DEBUG oslo_vmware.api [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465124, 'name': PowerOffVM_Task, 'duration_secs': 0.166214} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.781639] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1230.781727] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1230.782031] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ee80baf0-3c0d-4f2a-9661-d16180a6fed5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.874951] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1230.875199] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1230.875385] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleting the datastore file [datastore1] fb1195b2-1ab5-47d7-be86-4b61e47b4e29 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1230.875671] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c753f5d-adbb-47d0-a02a-837e60c2c751 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.882540] env[61728]: DEBUG oslo_vmware.api [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1230.882540] env[61728]: value = "task-465126" [ 1230.882540] env[61728]: _type = "Task" [ 1230.882540] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.890033] env[61728]: DEBUG oslo_vmware.api [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465126, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.975569] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52c5381a-2228-ff56-312c-b1940becf742, 'name': SearchDatastore_Task, 'duration_secs': 0.010501} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.976320] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1230.976692] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] f95e28ce-0792-488a-a730-0766272ab2d5/f95e28ce-0792-488a-a730-0766272ab2d5.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1230.977080] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a7c08a42-52b0-4e42-9e1f-b33b894f04df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.984139] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1230.984139] env[61728]: value = "task-465127" [ 1230.984139] env[61728]: _type = "Task" [ 1230.984139] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.993659] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465127, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.392472] env[61728]: DEBUG oslo_vmware.api [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465126, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.412513} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.392738] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1231.392930] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1231.393133] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1231.393314] env[61728]: INFO nova.compute.manager [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1231.393562] env[61728]: DEBUG oslo.service.loopingcall [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1231.393768] env[61728]: DEBUG nova.compute.manager [-] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1231.393870] env[61728]: DEBUG nova.network.neutron [-] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1231.494590] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465127, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.808350] env[61728]: DEBUG nova.compute.manager [req-7c6e1b9a-09d9-4133-9c27-eae425b8769c req-6af86795-0e33-40ea-95ca-5fba877e1ece service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Received event network-vif-deleted-8787e957-2e8a-4cce-b359-84b4b077d388 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1231.808350] env[61728]: INFO nova.compute.manager [req-7c6e1b9a-09d9-4133-9c27-eae425b8769c req-6af86795-0e33-40ea-95ca-5fba877e1ece service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Neutron deleted interface 8787e957-2e8a-4cce-b359-84b4b077d388; detaching it from the instance and deleting it from the info cache [ 1231.808350] env[61728]: DEBUG nova.network.neutron [req-7c6e1b9a-09d9-4133-9c27-eae425b8769c req-6af86795-0e33-40ea-95ca-5fba877e1ece service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1231.995355] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465127, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.571618} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.995600] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] f95e28ce-0792-488a-a730-0766272ab2d5/f95e28ce-0792-488a-a730-0766272ab2d5.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1231.995817] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1231.996085] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25da2a51-a6db-4ba5-aac1-9412e77f8ea1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.003504] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1232.003504] env[61728]: value = "task-465128" [ 1232.003504] env[61728]: _type = "Task" [ 1232.003504] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.011281] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465128, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.287296] env[61728]: DEBUG nova.network.neutron [-] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.312049] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8339e725-e832-4b2f-bc08-c846253a1019 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.321597] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54fa1fd-705c-4075-b60e-f8ee4995adfc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.346412] env[61728]: DEBUG nova.compute.manager [req-7c6e1b9a-09d9-4133-9c27-eae425b8769c req-6af86795-0e33-40ea-95ca-5fba877e1ece service nova] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Detach interface failed, port_id=8787e957-2e8a-4cce-b359-84b4b077d388, reason: Instance fb1195b2-1ab5-47d7-be86-4b61e47b4e29 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1232.514106] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465128, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089652} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.514106] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1232.514446] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d606c89-6e63-427c-9fc2-c98f35132fee {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.535120] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Reconfiguring VM instance instance-0000007c to attach disk [datastore1] f95e28ce-0792-488a-a730-0766272ab2d5/f95e28ce-0792-488a-a730-0766272ab2d5.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1232.535334] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-673d616e-cc9a-4698-9acd-372fbff6f661 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.554462] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1232.554462] env[61728]: value = "task-465129" [ 1232.554462] env[61728]: _type = "Task" [ 1232.554462] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.561897] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465129, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.790311] env[61728]: INFO nova.compute.manager [-] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Took 1.40 seconds to deallocate network for instance. [ 1233.064233] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465129, 'name': ReconfigVM_Task, 'duration_secs': 0.396397} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.064494] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Reconfigured VM instance instance-0000007c to attach disk [datastore1] f95e28ce-0792-488a-a730-0766272ab2d5/f95e28ce-0792-488a-a730-0766272ab2d5.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1233.065134] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0810e76-9695-4959-b20a-5d6399d3b297 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.071352] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1233.071352] env[61728]: value = "task-465130" [ 1233.071352] env[61728]: _type = "Task" [ 1233.071352] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.078766] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465130, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.297222] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1233.297543] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1233.297744] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1233.318496] env[61728]: INFO nova.scheduler.client.report [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleted allocations for instance fb1195b2-1ab5-47d7-be86-4b61e47b4e29 [ 1233.581446] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465130, 'name': Rename_Task, 'duration_secs': 0.170634} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.581777] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1233.581955] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32562eca-2da3-4b26-b82a-b06ecdc7808b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.587462] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1233.587462] env[61728]: value = "task-465131" [ 1233.587462] env[61728]: _type = "Task" [ 1233.587462] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.594464] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465131, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.826604] env[61728]: DEBUG oslo_concurrency.lockutils [None req-6de3eae1-9648-40e3-a077-d0dd0ae550b1 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "fb1195b2-1ab5-47d7-be86-4b61e47b4e29" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.575s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.097999] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465131, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.598217] env[61728]: DEBUG oslo_vmware.api [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465131, 'name': PowerOnVM_Task, 'duration_secs': 0.674039} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.598499] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1234.598704] env[61728]: INFO nova.compute.manager [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Took 9.61 seconds to spawn the instance on the hypervisor. [ 1234.598888] env[61728]: DEBUG nova.compute.manager [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1234.599665] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f660c11-ee3a-44fc-bb65-eace61e8661c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.116081] env[61728]: INFO nova.compute.manager [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Took 15.69 seconds to build instance. [ 1235.185282] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "37fe7d40-cce0-41c2-be6f-d608f3503077" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1235.185512] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "37fe7d40-cce0-41c2-be6f-d608f3503077" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1235.617880] env[61728]: DEBUG oslo_concurrency.lockutils [None req-5819ce19-0e2b-4ffd-bb5a-4258fe568135 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "f95e28ce-0792-488a-a730-0766272ab2d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.196s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1235.687606] env[61728]: DEBUG nova.compute.manager [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Starting instance... {{(pid=61728) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1235.945425] env[61728]: DEBUG nova.compute.manager [req-72ce13a8-cea5-4b38-b359-0966d5bf3e08 req-17abda27-96dd-4cc2-bc35-6923e77ae509 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Received event network-changed-a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1235.945654] env[61728]: DEBUG nova.compute.manager [req-72ce13a8-cea5-4b38-b359-0966d5bf3e08 req-17abda27-96dd-4cc2-bc35-6923e77ae509 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Refreshing instance network info cache due to event network-changed-a3aff6d0-5474-4d96-9da8-d7b907af2138. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1235.945952] env[61728]: DEBUG oslo_concurrency.lockutils [req-72ce13a8-cea5-4b38-b359-0966d5bf3e08 req-17abda27-96dd-4cc2-bc35-6923e77ae509 service nova] Acquiring lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1235.946189] env[61728]: DEBUG oslo_concurrency.lockutils [req-72ce13a8-cea5-4b38-b359-0966d5bf3e08 req-17abda27-96dd-4cc2-bc35-6923e77ae509 service nova] Acquired lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1235.946403] env[61728]: DEBUG nova.network.neutron [req-72ce13a8-cea5-4b38-b359-0966d5bf3e08 req-17abda27-96dd-4cc2-bc35-6923e77ae509 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Refreshing network info cache for port a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1236.213543] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1236.213815] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1236.215470] env[61728]: INFO nova.compute.claims [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1236.642665] env[61728]: DEBUG nova.network.neutron [req-72ce13a8-cea5-4b38-b359-0966d5bf3e08 req-17abda27-96dd-4cc2-bc35-6923e77ae509 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updated VIF entry in instance network info cache for port a3aff6d0-5474-4d96-9da8-d7b907af2138. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1236.643052] env[61728]: DEBUG nova.network.neutron [req-72ce13a8-cea5-4b38-b359-0966d5bf3e08 req-17abda27-96dd-4cc2-bc35-6923e77ae509 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updating instance_info_cache with network_info: [{"id": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "address": "fa:16:3e:30:bd:13", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.237", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3aff6d0-54", "ovs_interfaceid": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1237.145990] env[61728]: DEBUG oslo_concurrency.lockutils [req-72ce13a8-cea5-4b38-b359-0966d5bf3e08 req-17abda27-96dd-4cc2-bc35-6923e77ae509 service nova] Releasing lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1237.276639] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691b7f07-8ce5-4b63-bf03-6acacbb1e4a6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.284196] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c74bae-88cc-4261-9ecd-ca087cf331ff {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.313167] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40db4a58-f6cd-440c-9cf4-402e2ff2db48 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.319850] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc1bf84-dd60-4d6c-bf19-9913ec1d172d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.332530] env[61728]: DEBUG nova.compute.provider_tree [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1237.835809] env[61728]: DEBUG nova.scheduler.client.report [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1238.342241] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.128s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.342731] env[61728]: DEBUG nova.compute.manager [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Start building networks asynchronously for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1238.848607] env[61728]: DEBUG nova.compute.utils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1238.851892] env[61728]: DEBUG nova.compute.manager [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Allocating IP information in the background. {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1238.852161] env[61728]: DEBUG nova.network.neutron [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] allocate_for_instance() {{(pid=61728) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1238.894696] env[61728]: DEBUG nova.policy [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67d7c6a6191045cf85f790acebaf811e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7f61c87df8b485b9a6370032dd7af37', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61728) authorize /opt/stack/nova/nova/policy.py:203}} [ 1239.171217] env[61728]: DEBUG nova.network.neutron [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Successfully created port: ff94ce10-67c4-4c4c-baf9-d236d1670542 {{(pid=61728) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1239.356100] env[61728]: DEBUG nova.compute.manager [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Start building block device mappings for instance. {{(pid=61728) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1240.366275] env[61728]: DEBUG nova.compute.manager [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Start spawning the instance on the hypervisor. {{(pid=61728) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1240.394926] env[61728]: DEBUG nova.virt.hardware [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-29T12:20:33Z,direct_url=,disk_format='vmdk',id=8b767102-1435-4827-a43b-8e2e25ec780b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fb11ba9be5014418bbf48b9cc32669bc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-29T12:20:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1240.395200] env[61728]: DEBUG nova.virt.hardware [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1240.395363] env[61728]: DEBUG nova.virt.hardware [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1240.395551] env[61728]: DEBUG nova.virt.hardware [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1240.395702] env[61728]: DEBUG nova.virt.hardware [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1240.395855] env[61728]: DEBUG nova.virt.hardware [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1240.396083] env[61728]: DEBUG nova.virt.hardware [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1240.396256] env[61728]: DEBUG nova.virt.hardware [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1240.396430] env[61728]: DEBUG nova.virt.hardware [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1240.396598] env[61728]: DEBUG nova.virt.hardware [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1240.396775] env[61728]: DEBUG nova.virt.hardware [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1240.397674] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78efd02b-4fed-4820-ba81-56548a722df8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.405705] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8bd90a-1c3b-489b-9e51-9c34d88ac38c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.527073] env[61728]: DEBUG nova.compute.manager [req-79b5a436-b987-435f-b51c-d84ecd788504 req-c1f47c27-82b3-4066-944b-bd76203eb4a1 service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Received event network-vif-plugged-ff94ce10-67c4-4c4c-baf9-d236d1670542 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1240.527150] env[61728]: DEBUG oslo_concurrency.lockutils [req-79b5a436-b987-435f-b51c-d84ecd788504 req-c1f47c27-82b3-4066-944b-bd76203eb4a1 service nova] Acquiring lock "37fe7d40-cce0-41c2-be6f-d608f3503077-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.527358] env[61728]: DEBUG oslo_concurrency.lockutils [req-79b5a436-b987-435f-b51c-d84ecd788504 req-c1f47c27-82b3-4066-944b-bd76203eb4a1 service nova] Lock "37fe7d40-cce0-41c2-be6f-d608f3503077-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.527541] env[61728]: DEBUG oslo_concurrency.lockutils [req-79b5a436-b987-435f-b51c-d84ecd788504 req-c1f47c27-82b3-4066-944b-bd76203eb4a1 service nova] Lock "37fe7d40-cce0-41c2-be6f-d608f3503077-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.527712] env[61728]: DEBUG nova.compute.manager [req-79b5a436-b987-435f-b51c-d84ecd788504 req-c1f47c27-82b3-4066-944b-bd76203eb4a1 service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] No waiting events found dispatching network-vif-plugged-ff94ce10-67c4-4c4c-baf9-d236d1670542 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1240.527881] env[61728]: WARNING nova.compute.manager [req-79b5a436-b987-435f-b51c-d84ecd788504 req-c1f47c27-82b3-4066-944b-bd76203eb4a1 service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Received unexpected event network-vif-plugged-ff94ce10-67c4-4c4c-baf9-d236d1670542 for instance with vm_state building and task_state spawning. [ 1241.047397] env[61728]: DEBUG nova.network.neutron [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Successfully updated port: ff94ce10-67c4-4c4c-baf9-d236d1670542 {{(pid=61728) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1241.072010] env[61728]: DEBUG nova.compute.manager [req-03df8597-22a2-40ce-99d8-56cd5de2332a req-5b87a612-bd13-44bb-902c-6180dff778b0 service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Received event network-changed-ff94ce10-67c4-4c4c-baf9-d236d1670542 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1241.072010] env[61728]: DEBUG nova.compute.manager [req-03df8597-22a2-40ce-99d8-56cd5de2332a req-5b87a612-bd13-44bb-902c-6180dff778b0 service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Refreshing instance network info cache due to event network-changed-ff94ce10-67c4-4c4c-baf9-d236d1670542. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1241.072010] env[61728]: DEBUG oslo_concurrency.lockutils [req-03df8597-22a2-40ce-99d8-56cd5de2332a req-5b87a612-bd13-44bb-902c-6180dff778b0 service nova] Acquiring lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1241.072010] env[61728]: DEBUG oslo_concurrency.lockutils [req-03df8597-22a2-40ce-99d8-56cd5de2332a req-5b87a612-bd13-44bb-902c-6180dff778b0 service nova] Acquired lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.072157] env[61728]: DEBUG nova.network.neutron [req-03df8597-22a2-40ce-99d8-56cd5de2332a req-5b87a612-bd13-44bb-902c-6180dff778b0 service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Refreshing network info cache for port ff94ce10-67c4-4c4c-baf9-d236d1670542 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1241.549967] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1241.602532] env[61728]: DEBUG nova.network.neutron [req-03df8597-22a2-40ce-99d8-56cd5de2332a req-5b87a612-bd13-44bb-902c-6180dff778b0 service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1241.670084] env[61728]: DEBUG nova.network.neutron [req-03df8597-22a2-40ce-99d8-56cd5de2332a req-5b87a612-bd13-44bb-902c-6180dff778b0 service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.173357] env[61728]: DEBUG oslo_concurrency.lockutils [req-03df8597-22a2-40ce-99d8-56cd5de2332a req-5b87a612-bd13-44bb-902c-6180dff778b0 service nova] Releasing lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1242.173886] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.174140] env[61728]: DEBUG nova.network.neutron [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1242.704137] env[61728]: DEBUG nova.network.neutron [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Instance cache missing network info. {{(pid=61728) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1242.848099] env[61728]: DEBUG nova.network.neutron [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Updating instance_info_cache with network_info: [{"id": "ff94ce10-67c4-4c4c-baf9-d236d1670542", "address": "fa:16:3e:3d:06:bf", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff94ce10-67", "ovs_interfaceid": "ff94ce10-67c4-4c4c-baf9-d236d1670542", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1243.350892] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1243.351267] env[61728]: DEBUG nova.compute.manager [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Instance network_info: |[{"id": "ff94ce10-67c4-4c4c-baf9-d236d1670542", "address": "fa:16:3e:3d:06:bf", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff94ce10-67", "ovs_interfaceid": "ff94ce10-67c4-4c4c-baf9-d236d1670542", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61728) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1243.351722] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:06:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b80dd748-3d7e-4a23-a38d-9e79a3881452', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff94ce10-67c4-4c4c-baf9-d236d1670542', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1243.359145] env[61728]: DEBUG oslo.service.loopingcall [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1243.359368] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1243.359597] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-702caedc-52e6-452d-8968-43897e839302 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.380976] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1243.380976] env[61728]: value = "task-465132" [ 1243.380976] env[61728]: _type = "Task" [ 1243.380976] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.388440] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465132, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.890688] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465132, 'name': CreateVM_Task, 'duration_secs': 0.403208} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.890954] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1243.891584] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1243.891760] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.892121] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1243.892381] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a73a240-3a0b-43bc-b644-6241413b03d3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.896730] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1243.896730] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52005b3a-c0d9-bcb3-1089-afeaedd1fb8c" [ 1243.896730] env[61728]: _type = "Task" [ 1243.896730] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.904273] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52005b3a-c0d9-bcb3-1089-afeaedd1fb8c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.933884] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1243.933884] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1244.406929] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52005b3a-c0d9-bcb3-1089-afeaedd1fb8c, 'name': SearchDatastore_Task, 'duration_secs': 0.009761} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.407269] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1244.407563] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Processing image 8b767102-1435-4827-a43b-8e2e25ec780b {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1244.407806] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1244.407960] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.408161] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1244.408427] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0544c6d4-6966-4a20-bddc-e7882f9c1eb7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.416275] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1244.416455] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1244.417135] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3e4edd7-fda4-49f0-98bf-80c68f53bde6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.421900] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1244.421900] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52585379-0dea-fce3-c9ba-815dcbfcbcf3" [ 1244.421900] env[61728]: _type = "Task" [ 1244.421900] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.429117] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52585379-0dea-fce3-c9ba-815dcbfcbcf3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.436128] env[61728]: DEBUG nova.compute.utils [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1244.932986] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52585379-0dea-fce3-c9ba-815dcbfcbcf3, 'name': SearchDatastore_Task, 'duration_secs': 0.010608} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.933742] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea20bf53-5029-4445-8f8d-a4ef3c76f256 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.938900] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.939468] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1244.939468] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528178e5-2d05-5797-d4eb-43469b06a58c" [ 1244.939468] env[61728]: _type = "Task" [ 1244.939468] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.948793] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528178e5-2d05-5797-d4eb-43469b06a58c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.449634] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]528178e5-2d05-5797-d4eb-43469b06a58c, 'name': SearchDatastore_Task, 'duration_secs': 0.010828} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.450017] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1245.450215] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 37fe7d40-cce0-41c2-be6f-d608f3503077/37fe7d40-cce0-41c2-be6f-d608f3503077.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1245.450483] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-239084eb-9232-41e2-91ca-e53737c3f82e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.457283] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1245.457283] env[61728]: value = "task-465133" [ 1245.457283] env[61728]: _type = "Task" [ 1245.457283] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.464969] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465133, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.967480] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465133, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475907} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.967849] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b767102-1435-4827-a43b-8e2e25ec780b/8b767102-1435-4827-a43b-8e2e25ec780b.vmdk to [datastore1] 37fe7d40-cce0-41c2-be6f-d608f3503077/37fe7d40-cce0-41c2-be6f-d608f3503077.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1245.967950] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Extending root virtual disk to 1048576 {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1245.968222] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d5478780-5184-4038-a2c7-ddc9b343c211 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.974366] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1245.974366] env[61728]: value = "task-465134" [ 1245.974366] env[61728]: _type = "Task" [ 1245.974366] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.981466] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465134, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.000283] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1246.000524] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1246.000768] env[61728]: INFO nova.compute.manager [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Attaching volume a9266eb4-98f9-40e7-9b5c-2c3717f94c33 to /dev/sdb [ 1246.033234] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4a2b29-d58e-4d00-bc51-369be672003d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.040713] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2230a69f-4e3c-4580-ba3f-7a46d0e5f07d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.055905] env[61728]: DEBUG nova.virt.block_device [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Updating existing volume attachment record: c798fa68-6098-4c55-8fa8-bb986b846085 {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1246.484526] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465134, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071285} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.484800] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Extended root virtual disk {{(pid=61728) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1246.485597] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ff96b9-636b-41c9-b43a-83a5c130cda6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.507022] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Reconfiguring VM instance instance-0000007d to attach disk [datastore1] 37fe7d40-cce0-41c2-be6f-d608f3503077/37fe7d40-cce0-41c2-be6f-d608f3503077.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1246.507319] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb0355b4-48c4-40e2-87c6-867009c49a95 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.526371] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1246.526371] env[61728]: value = "task-465136" [ 1246.526371] env[61728]: _type = "Task" [ 1246.526371] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.534506] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465136, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.036718] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465136, 'name': ReconfigVM_Task, 'duration_secs': 0.296494} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.037090] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Reconfigured VM instance instance-0000007d to attach disk [datastore1] 37fe7d40-cce0-41c2-be6f-d608f3503077/37fe7d40-cce0-41c2-be6f-d608f3503077.vmdk or device None with type sparse {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1247.037674] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad66e9a4-a314-4a3a-b37f-f9266ab79c7e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.045020] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1247.045020] env[61728]: value = "task-465137" [ 1247.045020] env[61728]: _type = "Task" [ 1247.045020] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.052298] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465137, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.555429] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465137, 'name': Rename_Task, 'duration_secs': 0.164005} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.555709] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1247.555952] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-16623538-c336-452b-bae2-0a169647b290 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.562429] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1247.562429] env[61728]: value = "task-465138" [ 1247.562429] env[61728]: _type = "Task" [ 1247.562429] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.569685] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465138, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.842836] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1247.843090] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1247.843265] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Starting heal instance info cache {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1247.843404] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Rebuilding the list of instances to heal {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1248.072553] env[61728]: DEBUG oslo_vmware.api [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465138, 'name': PowerOnVM_Task, 'duration_secs': 0.470283} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.072922] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1248.073044] env[61728]: INFO nova.compute.manager [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Took 7.71 seconds to spawn the instance on the hypervisor. [ 1248.073237] env[61728]: DEBUG nova.compute.manager [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1248.074061] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db1c66b-9fbd-4006-bbcb-1bfbfa458dc7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.347132] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Skipping network cache update for instance because it is Building. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1248.371805] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "refresh_cache-e5943edd-bdc3-4285-b2e0-7adac9acaf7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1248.371969] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquired lock "refresh_cache-e5943edd-bdc3-4285-b2e0-7adac9acaf7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1248.372130] env[61728]: DEBUG nova.network.neutron [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Forcefully refreshing network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1248.372285] env[61728]: DEBUG nova.objects.instance [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lazy-loading 'info_cache' on Instance uuid e5943edd-bdc3-4285-b2e0-7adac9acaf7f {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1248.591624] env[61728]: INFO nova.compute.manager [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Took 12.40 seconds to build instance. [ 1249.093793] env[61728]: DEBUG oslo_concurrency.lockutils [None req-286d6f3e-5219-45a7-9cd7-f96715e6c7fa tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "37fe7d40-cce0-41c2-be6f-d608f3503077" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.908s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.317937] env[61728]: DEBUG nova.compute.manager [req-333c4172-7a38-49d0-a4a9-b0406324e982 req-21b0edf7-544b-4a25-8c0a-a0fb4c6d84ad service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Received event network-changed-ff94ce10-67c4-4c4c-baf9-d236d1670542 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1249.318174] env[61728]: DEBUG nova.compute.manager [req-333c4172-7a38-49d0-a4a9-b0406324e982 req-21b0edf7-544b-4a25-8c0a-a0fb4c6d84ad service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Refreshing instance network info cache due to event network-changed-ff94ce10-67c4-4c4c-baf9-d236d1670542. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1249.318364] env[61728]: DEBUG oslo_concurrency.lockutils [req-333c4172-7a38-49d0-a4a9-b0406324e982 req-21b0edf7-544b-4a25-8c0a-a0fb4c6d84ad service nova] Acquiring lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1249.318518] env[61728]: DEBUG oslo_concurrency.lockutils [req-333c4172-7a38-49d0-a4a9-b0406324e982 req-21b0edf7-544b-4a25-8c0a-a0fb4c6d84ad service nova] Acquired lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1249.318684] env[61728]: DEBUG nova.network.neutron [req-333c4172-7a38-49d0-a4a9-b0406324e982 req-21b0edf7-544b-4a25-8c0a-a0fb4c6d84ad service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Refreshing network info cache for port ff94ce10-67c4-4c4c-baf9-d236d1670542 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1250.012672] env[61728]: DEBUG nova.network.neutron [req-333c4172-7a38-49d0-a4a9-b0406324e982 req-21b0edf7-544b-4a25-8c0a-a0fb4c6d84ad service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Updated VIF entry in instance network info cache for port ff94ce10-67c4-4c4c-baf9-d236d1670542. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1250.013068] env[61728]: DEBUG nova.network.neutron [req-333c4172-7a38-49d0-a4a9-b0406324e982 req-21b0edf7-544b-4a25-8c0a-a0fb4c6d84ad service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Updating instance_info_cache with network_info: [{"id": "ff94ce10-67c4-4c4c-baf9-d236d1670542", "address": "fa:16:3e:3d:06:bf", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff94ce10-67", "ovs_interfaceid": "ff94ce10-67c4-4c4c-baf9-d236d1670542", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1250.079014] env[61728]: DEBUG nova.network.neutron [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Updating instance_info_cache with network_info: [{"id": "e4ebecc4-1e02-4afe-aa2f-25362f481dc0", "address": "fa:16:3e:84:9c:d5", "network": {"id": "72d11e2a-907e-419e-b8a4-53aa39f2566a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1755247352-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833a646ea21a48c4951ed8cedebaa98f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3836fb52-19c6-4e10-a0ca-f0bca73dc887", "external-id": "nsx-vlan-transportzone-964", "segmentation_id": 964, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4ebecc4-1e", "ovs_interfaceid": "e4ebecc4-1e02-4afe-aa2f-25362f481dc0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1250.515823] env[61728]: DEBUG oslo_concurrency.lockutils [req-333c4172-7a38-49d0-a4a9-b0406324e982 req-21b0edf7-544b-4a25-8c0a-a0fb4c6d84ad service nova] Releasing lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1250.581820] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Releasing lock "refresh_cache-e5943edd-bdc3-4285-b2e0-7adac9acaf7f" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1250.582032] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Updated the network info_cache for instance {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1250.582204] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1250.582366] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1250.582513] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1250.582661] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1250.582812] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1250.582985] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1250.583152] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61728) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1250.583303] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1250.602499] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1250.602793] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122246', 'volume_id': 'a9266eb4-98f9-40e7-9b5c-2c3717f94c33', 'name': 'volume-a9266eb4-98f9-40e7-9b5c-2c3717f94c33', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e5943edd-bdc3-4285-b2e0-7adac9acaf7f', 'attached_at': '', 'detached_at': '', 'volume_id': 'a9266eb4-98f9-40e7-9b5c-2c3717f94c33', 'serial': 'a9266eb4-98f9-40e7-9b5c-2c3717f94c33'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1250.604082] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af1d610-f09b-452c-99dd-e94929694ae9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.622609] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb4297b-f219-4039-b015-3db6534e5a86 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.649179] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Reconfiguring VM instance instance-0000007b to attach disk [datastore1] volume-a9266eb4-98f9-40e7-9b5c-2c3717f94c33/volume-a9266eb4-98f9-40e7-9b5c-2c3717f94c33.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1250.649477] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c8e1f65-b3ba-47d0-a6b1-ddd1605d43f3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.667923] env[61728]: DEBUG oslo_vmware.api [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1250.667923] env[61728]: value = "task-465140" [ 1250.667923] env[61728]: _type = "Task" [ 1250.667923] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.679679] env[61728]: DEBUG oslo_vmware.api [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465140, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.087434] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1251.087728] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1251.087825] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1251.087981] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61728) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1251.088896] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759d7d60-741d-4742-affb-2ddbb9b03306 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.097130] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54db216b-7c55-49ed-a09d-d9cfda04e7b6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.111267] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e432a6de-327f-4d3e-9f51-5d05a3ed381a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.117962] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7242c1bb-3328-405e-bac3-6263e3fcd303 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.147604] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181387MB free_disk=115GB free_vcpus=48 pci_devices=None {{(pid=61728) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1251.147785] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1251.147936] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1251.179098] env[61728]: DEBUG oslo_vmware.api [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465140, 'name': ReconfigVM_Task, 'duration_secs': 0.402183} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.179349] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Reconfigured VM instance instance-0000007b to attach disk [datastore1] volume-a9266eb4-98f9-40e7-9b5c-2c3717f94c33/volume-a9266eb4-98f9-40e7-9b5c-2c3717f94c33.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1251.184294] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69b54d97-79bb-4487-a51e-24ab16e34b98 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.199349] env[61728]: DEBUG oslo_vmware.api [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1251.199349] env[61728]: value = "task-465141" [ 1251.199349] env[61728]: _type = "Task" [ 1251.199349] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.207245] env[61728]: DEBUG oslo_vmware.api [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465141, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.710247] env[61728]: DEBUG oslo_vmware.api [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465141, 'name': ReconfigVM_Task, 'duration_secs': 0.15684} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.710566] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122246', 'volume_id': 'a9266eb4-98f9-40e7-9b5c-2c3717f94c33', 'name': 'volume-a9266eb4-98f9-40e7-9b5c-2c3717f94c33', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e5943edd-bdc3-4285-b2e0-7adac9acaf7f', 'attached_at': '', 'detached_at': '', 'volume_id': 'a9266eb4-98f9-40e7-9b5c-2c3717f94c33', 'serial': 'a9266eb4-98f9-40e7-9b5c-2c3717f94c33'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1252.174755] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance e5943edd-bdc3-4285-b2e0-7adac9acaf7f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1252.174755] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance f95e28ce-0792-488a-a730-0766272ab2d5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1252.174755] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance 37fe7d40-cce0-41c2-be6f-d608f3503077 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1252.174755] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1252.174755] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1252.226495] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5506303d-83be-4a10-8ba7-f66bb5a4cc91 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.234495] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7abdb2ef-d7ee-4abc-9987-36292bbfca7a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.267163] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b65e926-89a5-4047-8a09-2ddd3f864a87 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.274736] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910bb807-5f5d-46b7-8a65-83112eb918fe {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.289171] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1252.772032] env[61728]: DEBUG nova.objects.instance [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lazy-loading 'flavor' on Instance uuid e5943edd-bdc3-4285-b2e0-7adac9acaf7f {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1252.792369] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1253.276886] env[61728]: DEBUG oslo_concurrency.lockutils [None req-b8484c1a-6b2a-4544-b3d2-e2ddaca75a32 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.276s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1253.296881] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1253.297086] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.149s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1254.068029] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.068029] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.571251] env[61728]: DEBUG nova.compute.utils [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Using /dev/sd instead of None {{(pid=61728) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1255.074464] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1256.132100] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1256.132444] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1256.132648] env[61728]: INFO nova.compute.manager [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Attaching volume a74edc62-e926-44eb-87b7-811798282879 to /dev/sdc [ 1256.162847] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e495f6-36b5-4dd0-9d3c-ce6643bcff56 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.170219] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fac265c-d2ec-49c3-bf98-6ce78802128e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.182692] env[61728]: DEBUG nova.virt.block_device [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Updating existing volume attachment record: 8fffc0df-9fb8-4086-be85-030bdadc0cf8 {{(pid=61728) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1260.726994] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Volume attach. Driver type: vmdk {{(pid=61728) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1260.727320] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122247', 'volume_id': 'a74edc62-e926-44eb-87b7-811798282879', 'name': 'volume-a74edc62-e926-44eb-87b7-811798282879', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e5943edd-bdc3-4285-b2e0-7adac9acaf7f', 'attached_at': '', 'detached_at': '', 'volume_id': 'a74edc62-e926-44eb-87b7-811798282879', 'serial': 'a74edc62-e926-44eb-87b7-811798282879'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1260.728221] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3545b5de-fd0f-41b3-965b-b6c85db02179 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.744204] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5432ccc5-720e-4d7e-9c8d-016a5cb7d121 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.769821] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Reconfiguring VM instance instance-0000007b to attach disk [datastore1] volume-a74edc62-e926-44eb-87b7-811798282879/volume-a74edc62-e926-44eb-87b7-811798282879.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1260.770064] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bad85bf5-8bf0-4a9e-8474-bd2f68130080 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.787413] env[61728]: DEBUG oslo_vmware.api [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1260.787413] env[61728]: value = "task-465144" [ 1260.787413] env[61728]: _type = "Task" [ 1260.787413] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.794650] env[61728]: DEBUG oslo_vmware.api [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465144, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.299460] env[61728]: DEBUG oslo_vmware.api [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465144, 'name': ReconfigVM_Task, 'duration_secs': 0.369991} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.299861] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Reconfigured VM instance instance-0000007b to attach disk [datastore1] volume-a74edc62-e926-44eb-87b7-811798282879/volume-a74edc62-e926-44eb-87b7-811798282879.vmdk or device None with type thin {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1261.304612] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81c27cd1-0515-4e33-a283-8a3b4d419053 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.319168] env[61728]: DEBUG oslo_vmware.api [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1261.319168] env[61728]: value = "task-465145" [ 1261.319168] env[61728]: _type = "Task" [ 1261.319168] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.327204] env[61728]: DEBUG oslo_vmware.api [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465145, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.829292] env[61728]: DEBUG oslo_vmware.api [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465145, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.329934] env[61728]: DEBUG oslo_vmware.api [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465145, 'name': ReconfigVM_Task, 'duration_secs': 0.926745} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.330249] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122247', 'volume_id': 'a74edc62-e926-44eb-87b7-811798282879', 'name': 'volume-a74edc62-e926-44eb-87b7-811798282879', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e5943edd-bdc3-4285-b2e0-7adac9acaf7f', 'attached_at': '', 'detached_at': '', 'volume_id': 'a74edc62-e926-44eb-87b7-811798282879', 'serial': 'a74edc62-e926-44eb-87b7-811798282879'} {{(pid=61728) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1263.365350] env[61728]: DEBUG nova.objects.instance [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lazy-loading 'flavor' on Instance uuid e5943edd-bdc3-4285-b2e0-7adac9acaf7f {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1263.872490] env[61728]: DEBUG oslo_concurrency.lockutils [None req-fc536bda-434d-42d4-b04c-f7cbb40b501f tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.740s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1264.681212] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1264.681505] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1265.185943] env[61728]: INFO nova.compute.manager [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Detaching volume a9266eb4-98f9-40e7-9b5c-2c3717f94c33 [ 1265.216457] env[61728]: INFO nova.virt.block_device [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Attempting to driver detach volume a9266eb4-98f9-40e7-9b5c-2c3717f94c33 from mountpoint /dev/sdb [ 1265.216707] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1265.216899] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122246', 'volume_id': 'a9266eb4-98f9-40e7-9b5c-2c3717f94c33', 'name': 'volume-a9266eb4-98f9-40e7-9b5c-2c3717f94c33', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e5943edd-bdc3-4285-b2e0-7adac9acaf7f', 'attached_at': '', 'detached_at': '', 'volume_id': 'a9266eb4-98f9-40e7-9b5c-2c3717f94c33', 'serial': 'a9266eb4-98f9-40e7-9b5c-2c3717f94c33'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1265.217796] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f84fed2d-76e2-4407-9bcd-a6366b475cfc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.242601] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11fbe34-1146-45fb-bdd5-84477a8bd76f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.249029] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d500d8-d8d0-458d-9db7-134a2ac58b4f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.270944] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18662b15-45cf-43da-9811-06c0998d8a22 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.284942] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] The volume has not been displaced from its original location: [datastore1] volume-a9266eb4-98f9-40e7-9b5c-2c3717f94c33/volume-a9266eb4-98f9-40e7-9b5c-2c3717f94c33.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1265.290108] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Reconfiguring VM instance instance-0000007b to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1265.290372] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-70459913-aacd-4086-b7dc-344d6613a1da {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.307487] env[61728]: DEBUG oslo_vmware.api [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1265.307487] env[61728]: value = "task-465146" [ 1265.307487] env[61728]: _type = "Task" [ 1265.307487] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.318431] env[61728]: DEBUG oslo_vmware.api [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465146, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.816720] env[61728]: DEBUG oslo_vmware.api [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465146, 'name': ReconfigVM_Task, 'duration_secs': 0.226741} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.817097] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Reconfigured VM instance instance-0000007b to detach disk 2001 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1265.821624] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d83a5b10-ec90-4373-8a86-f3387a6c7ba2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.835849] env[61728]: DEBUG oslo_vmware.api [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1265.835849] env[61728]: value = "task-465147" [ 1265.835849] env[61728]: _type = "Task" [ 1265.835849] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.843268] env[61728]: DEBUG oslo_vmware.api [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465147, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.345894] env[61728]: DEBUG oslo_vmware.api [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465147, 'name': ReconfigVM_Task, 'duration_secs': 0.185442} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.346231] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122246', 'volume_id': 'a9266eb4-98f9-40e7-9b5c-2c3717f94c33', 'name': 'volume-a9266eb4-98f9-40e7-9b5c-2c3717f94c33', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e5943edd-bdc3-4285-b2e0-7adac9acaf7f', 'attached_at': '', 'detached_at': '', 'volume_id': 'a9266eb4-98f9-40e7-9b5c-2c3717f94c33', 'serial': 'a9266eb4-98f9-40e7-9b5c-2c3717f94c33'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1266.889920] env[61728]: DEBUG nova.objects.instance [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lazy-loading 'flavor' on Instance uuid e5943edd-bdc3-4285-b2e0-7adac9acaf7f {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1267.897629] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c93ec3cd-73b7-4a5a-9968-8b0dbd856cff tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.216s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1267.913178] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1267.913424] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.416079] env[61728]: INFO nova.compute.manager [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Detaching volume a74edc62-e926-44eb-87b7-811798282879 [ 1268.445686] env[61728]: INFO nova.virt.block_device [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Attempting to driver detach volume a74edc62-e926-44eb-87b7-811798282879 from mountpoint /dev/sdc [ 1268.445925] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Volume detach. Driver type: vmdk {{(pid=61728) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1268.446131] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122247', 'volume_id': 'a74edc62-e926-44eb-87b7-811798282879', 'name': 'volume-a74edc62-e926-44eb-87b7-811798282879', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e5943edd-bdc3-4285-b2e0-7adac9acaf7f', 'attached_at': '', 'detached_at': '', 'volume_id': 'a74edc62-e926-44eb-87b7-811798282879', 'serial': 'a74edc62-e926-44eb-87b7-811798282879'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1268.446981] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523e0c5c-088a-483f-8203-e0706e30d873 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.467629] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24c7c9af-5cf5-4804-bf1d-5ecbdb1d052d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.473931] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b81ef15-8c87-453c-88fe-6c5966d3abe7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.492564] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c0ffdd-8ea3-4cad-b1bd-9598d086404a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.506582] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] The volume has not been displaced from its original location: [datastore1] volume-a74edc62-e926-44eb-87b7-811798282879/volume-a74edc62-e926-44eb-87b7-811798282879.vmdk. No consolidation needed. {{(pid=61728) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1268.512210] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Reconfiguring VM instance instance-0000007b to detach disk 2002 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1268.512463] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c818938a-0891-4d19-8a48-07a5dc485438 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.529386] env[61728]: DEBUG oslo_vmware.api [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1268.529386] env[61728]: value = "task-465148" [ 1268.529386] env[61728]: _type = "Task" [ 1268.529386] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.536648] env[61728]: DEBUG oslo_vmware.api [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465148, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.039019] env[61728]: DEBUG oslo_vmware.api [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465148, 'name': ReconfigVM_Task, 'duration_secs': 0.220268} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.039371] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Reconfigured VM instance instance-0000007b to detach disk 2002 {{(pid=61728) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1269.044490] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c04735e2-1ef8-4277-b996-4e58378cc2da {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.059631] env[61728]: DEBUG oslo_vmware.api [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1269.059631] env[61728]: value = "task-465149" [ 1269.059631] env[61728]: _type = "Task" [ 1269.059631] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.068062] env[61728]: DEBUG oslo_vmware.api [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465149, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.569760] env[61728]: DEBUG oslo_vmware.api [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465149, 'name': ReconfigVM_Task, 'duration_secs': 0.131037} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.570136] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-122247', 'volume_id': 'a74edc62-e926-44eb-87b7-811798282879', 'name': 'volume-a74edc62-e926-44eb-87b7-811798282879', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e5943edd-bdc3-4285-b2e0-7adac9acaf7f', 'attached_at': '', 'detached_at': '', 'volume_id': 'a74edc62-e926-44eb-87b7-811798282879', 'serial': 'a74edc62-e926-44eb-87b7-811798282879'} {{(pid=61728) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1270.114138] env[61728]: DEBUG nova.objects.instance [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lazy-loading 'flavor' on Instance uuid e5943edd-bdc3-4285-b2e0-7adac9acaf7f {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1271.122179] env[61728]: DEBUG oslo_concurrency.lockutils [None req-a8a056f8-405e-4f3d-b34b-68d4a27f86f5 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.208s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1272.274651] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1272.275053] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1272.275179] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1272.275374] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1272.275547] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1272.277884] env[61728]: INFO nova.compute.manager [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Terminating instance [ 1272.280076] env[61728]: DEBUG nova.compute.manager [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1272.280288] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1272.281172] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809d385b-53fa-4c2e-8eab-c965cb6847bc {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.290220] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1272.290452] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8cd54dbe-f046-440a-b347-987b68ad140d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.297349] env[61728]: DEBUG oslo_vmware.api [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1272.297349] env[61728]: value = "task-465150" [ 1272.297349] env[61728]: _type = "Task" [ 1272.297349] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.305031] env[61728]: DEBUG oslo_vmware.api [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465150, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.807318] env[61728]: DEBUG oslo_vmware.api [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465150, 'name': PowerOffVM_Task, 'duration_secs': 0.182196} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.807585] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1272.807759] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1272.808010] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57b95872-1203-4ce1-b769-1ad25c3a98c6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.873195] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1272.873425] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1272.873606] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Deleting the datastore file [datastore1] e5943edd-bdc3-4285-b2e0-7adac9acaf7f {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1272.873872] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cfa1fa51-f168-43d0-a5e1-1faa13301023 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.879350] env[61728]: DEBUG oslo_vmware.api [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for the task: (returnval){ [ 1272.879350] env[61728]: value = "task-465152" [ 1272.879350] env[61728]: _type = "Task" [ 1272.879350] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.886699] env[61728]: DEBUG oslo_vmware.api [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465152, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1273.389185] env[61728]: DEBUG oslo_vmware.api [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Task: {'id': task-465152, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155535} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1273.389609] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1273.389609] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1273.389755] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1273.389933] env[61728]: INFO nova.compute.manager [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1273.390200] env[61728]: DEBUG oslo.service.loopingcall [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1273.390405] env[61728]: DEBUG nova.compute.manager [-] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1273.390515] env[61728]: DEBUG nova.network.neutron [-] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1274.094718] env[61728]: DEBUG nova.compute.manager [req-35ba2e2c-a91b-4984-8935-5f34c127344d req-8a2b0118-fd3e-4c40-9cd9-f9e60587e7c8 service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Received event network-vif-deleted-e4ebecc4-1e02-4afe-aa2f-25362f481dc0 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1274.094942] env[61728]: INFO nova.compute.manager [req-35ba2e2c-a91b-4984-8935-5f34c127344d req-8a2b0118-fd3e-4c40-9cd9-f9e60587e7c8 service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Neutron deleted interface e4ebecc4-1e02-4afe-aa2f-25362f481dc0; detaching it from the instance and deleting it from the info cache [ 1274.095111] env[61728]: DEBUG nova.network.neutron [req-35ba2e2c-a91b-4984-8935-5f34c127344d req-8a2b0118-fd3e-4c40-9cd9-f9e60587e7c8 service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1274.274187] env[61728]: DEBUG nova.network.neutron [-] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1274.599862] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c428f3d-9af7-42ae-9f04-29250044ac9a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.609898] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df26885-24da-44b6-9a28-96c2ca07a63d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.636534] env[61728]: DEBUG nova.compute.manager [req-35ba2e2c-a91b-4984-8935-5f34c127344d req-8a2b0118-fd3e-4c40-9cd9-f9e60587e7c8 service nova] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Detach interface failed, port_id=e4ebecc4-1e02-4afe-aa2f-25362f481dc0, reason: Instance e5943edd-bdc3-4285-b2e0-7adac9acaf7f could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1274.776629] env[61728]: INFO nova.compute.manager [-] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Took 1.39 seconds to deallocate network for instance. [ 1275.283515] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1275.283515] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1275.283673] env[61728]: DEBUG nova.objects.instance [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lazy-loading 'resources' on Instance uuid e5943edd-bdc3-4285-b2e0-7adac9acaf7f {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1275.686375] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "f95e28ce-0792-488a-a730-0766272ab2d5" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1275.686778] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "f95e28ce-0792-488a-a730-0766272ab2d5" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1275.686891] env[61728]: INFO nova.compute.manager [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Shelving [ 1275.835234] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78cd4a52-400d-4b16-8d15-4f6a37240ec5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.843024] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c73d93d-97a4-4a24-ab0f-50ecd8d02c13 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.871512] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69952d4-0aaf-4f9d-8bf3-6291fa7f2399 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.878067] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba1d499-509f-4b46-9c66-4a509657583a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.891399] env[61728]: DEBUG nova.compute.provider_tree [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1276.194625] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1276.194994] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4cc21ccd-f691-4694-b147-f3202cf77e65 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.202329] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1276.202329] env[61728]: value = "task-465153" [ 1276.202329] env[61728]: _type = "Task" [ 1276.202329] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.211126] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465153, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.394227] env[61728]: DEBUG nova.scheduler.client.report [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1276.712472] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465153, 'name': PowerOffVM_Task, 'duration_secs': 0.2138} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.712828] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1276.713435] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff3751ce-a956-47a5-a1ed-75b1753ccfb3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.730411] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a37351-0420-49ba-b6be-82fdfeb1a631 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.898636] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.615s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1276.918071] env[61728]: INFO nova.scheduler.client.report [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Deleted allocations for instance e5943edd-bdc3-4285-b2e0-7adac9acaf7f [ 1277.239787] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Creating Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1277.240498] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7f37cb5a-300d-4b53-bff1-be71c2d6fe50 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.248708] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1277.248708] env[61728]: value = "task-465154" [ 1277.248708] env[61728]: _type = "Task" [ 1277.248708] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.256833] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465154, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.426754] env[61728]: DEBUG oslo_concurrency.lockutils [None req-89cb41a0-1db9-43e9-98cc-fad9616a38a4 tempest-AttachVolumeTestJSON-1293063066 tempest-AttachVolumeTestJSON-1293063066-project-member] Lock "e5943edd-bdc3-4285-b2e0-7adac9acaf7f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.152s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1277.758700] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465154, 'name': CreateSnapshot_Task, 'duration_secs': 0.465133} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.759104] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Created Snapshot of the VM instance {{(pid=61728) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1277.759834] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb33816-a84c-482f-8662-9d517693c49c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.279195] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Creating linked-clone VM from snapshot {{(pid=61728) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1278.279512] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c765d31a-a9e5-4e12-bc36-ab9913c19aa7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.288436] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1278.288436] env[61728]: value = "task-465156" [ 1278.288436] env[61728]: _type = "Task" [ 1278.288436] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.295864] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465156, 'name': CloneVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.798016] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465156, 'name': CloneVM_Task} progress is 94%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.298887] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465156, 'name': CloneVM_Task, 'duration_secs': 0.966579} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.299165] env[61728]: INFO nova.virt.vmwareapi.vmops [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Created linked-clone VM from snapshot [ 1279.299861] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52ae4a4-9dd9-45c9-998d-837fa47ff38d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.306866] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Uploading image 8a9a50dc-caf7-4cd3-afdb-3936725c0d38 {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1279.331105] env[61728]: DEBUG oslo_vmware.rw_handles [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1279.331105] env[61728]: value = "vm-122249" [ 1279.331105] env[61728]: _type = "VirtualMachine" [ 1279.331105] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1279.331356] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f8b6e032-7ede-4ab1-b760-e9b7a854ba53 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.338778] env[61728]: DEBUG oslo_vmware.rw_handles [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lease: (returnval){ [ 1279.338778] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5231d9ab-a354-aacd-dc6f-fe8ab086be8a" [ 1279.338778] env[61728]: _type = "HttpNfcLease" [ 1279.338778] env[61728]: } obtained for exporting VM: (result){ [ 1279.338778] env[61728]: value = "vm-122249" [ 1279.338778] env[61728]: _type = "VirtualMachine" [ 1279.338778] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1279.339045] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the lease: (returnval){ [ 1279.339045] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5231d9ab-a354-aacd-dc6f-fe8ab086be8a" [ 1279.339045] env[61728]: _type = "HttpNfcLease" [ 1279.339045] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1279.344746] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1279.344746] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5231d9ab-a354-aacd-dc6f-fe8ab086be8a" [ 1279.344746] env[61728]: _type = "HttpNfcLease" [ 1279.344746] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1279.847770] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1279.847770] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5231d9ab-a354-aacd-dc6f-fe8ab086be8a" [ 1279.847770] env[61728]: _type = "HttpNfcLease" [ 1279.847770] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1279.848305] env[61728]: DEBUG oslo_vmware.rw_handles [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1279.848305] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]5231d9ab-a354-aacd-dc6f-fe8ab086be8a" [ 1279.848305] env[61728]: _type = "HttpNfcLease" [ 1279.848305] env[61728]: }. {{(pid=61728) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1279.848799] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea158f6-d179-46ed-a029-0a0fd4bd6b78 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.856188] env[61728]: DEBUG oslo_vmware.rw_handles [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5218aced-8f91-51f7-7876-be4f3232957a/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1279.856390] env[61728]: DEBUG oslo_vmware.rw_handles [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5218aced-8f91-51f7-7876-be4f3232957a/disk-0.vmdk for reading. {{(pid=61728) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1279.941700] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b6d26784-3143-42a6-980e-522a690d5911 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.575584] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2a9f821f-55d9-4dd6-90fe-e527808c866b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "37fe7d40-cce0-41c2-be6f-d608f3503077" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1286.575974] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2a9f821f-55d9-4dd6-90fe-e527808c866b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "37fe7d40-cce0-41c2-be6f-d608f3503077" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1286.576080] env[61728]: DEBUG nova.compute.manager [None req-2a9f821f-55d9-4dd6-90fe-e527808c866b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1286.577041] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c19670-cff6-48bf-8226-10cb171e70da {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.583953] env[61728]: DEBUG nova.compute.manager [None req-2a9f821f-55d9-4dd6-90fe-e527808c866b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61728) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1286.584564] env[61728]: DEBUG nova.objects.instance [None req-2a9f821f-55d9-4dd6-90fe-e527808c866b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lazy-loading 'flavor' on Instance uuid 37fe7d40-cce0-41c2-be6f-d608f3503077 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1287.090539] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a9f821f-55d9-4dd6-90fe-e527808c866b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1287.090539] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6be2e34d-6272-4b22-8eee-37dc16cdf5a7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.098443] env[61728]: DEBUG oslo_vmware.api [None req-2a9f821f-55d9-4dd6-90fe-e527808c866b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1287.098443] env[61728]: value = "task-465160" [ 1287.098443] env[61728]: _type = "Task" [ 1287.098443] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.106993] env[61728]: DEBUG oslo_vmware.api [None req-2a9f821f-55d9-4dd6-90fe-e527808c866b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465160, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.610042] env[61728]: DEBUG oslo_vmware.api [None req-2a9f821f-55d9-4dd6-90fe-e527808c866b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465160, 'name': PowerOffVM_Task, 'duration_secs': 0.207683} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.610410] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a9f821f-55d9-4dd6-90fe-e527808c866b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1287.610640] env[61728]: DEBUG nova.compute.manager [None req-2a9f821f-55d9-4dd6-90fe-e527808c866b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1287.611474] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738695b1-5f5d-421e-b8e6-11a567121d57 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.951436] env[61728]: DEBUG oslo_vmware.rw_handles [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5218aced-8f91-51f7-7876-be4f3232957a/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1287.952469] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb7e581-2e69-4d73-a0a4-c76b3930a773 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.958742] env[61728]: DEBUG oslo_vmware.rw_handles [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5218aced-8f91-51f7-7876-be4f3232957a/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1287.958918] env[61728]: ERROR oslo_vmware.rw_handles [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5218aced-8f91-51f7-7876-be4f3232957a/disk-0.vmdk due to incomplete transfer. [ 1287.959155] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-acc73236-a65b-4b94-b2d8-d090b31fd674 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.966325] env[61728]: DEBUG oslo_vmware.rw_handles [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5218aced-8f91-51f7-7876-be4f3232957a/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1287.966525] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Uploaded image 8a9a50dc-caf7-4cd3-afdb-3936725c0d38 to the Glance image server {{(pid=61728) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1287.968936] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Destroying the VM {{(pid=61728) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1287.969173] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8209da7d-eb03-469c-86f2-ef0070f13871 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.974665] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1287.974665] env[61728]: value = "task-465161" [ 1287.974665] env[61728]: _type = "Task" [ 1287.974665] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.982519] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465161, 'name': Destroy_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.124320] env[61728]: DEBUG oslo_concurrency.lockutils [None req-2a9f821f-55d9-4dd6-90fe-e527808c866b tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "37fe7d40-cce0-41c2-be6f-d608f3503077" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.548s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1288.484025] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465161, 'name': Destroy_Task, 'duration_secs': 0.333056} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.484183] env[61728]: INFO nova.virt.vmwareapi.vm_util [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Destroyed the VM [ 1288.484423] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Deleting Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1288.484675] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-74edac63-d47a-4456-965f-e21c83be8d76 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.490773] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1288.490773] env[61728]: value = "task-465162" [ 1288.490773] env[61728]: _type = "Task" [ 1288.490773] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.498323] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465162, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.001188] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465162, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.024065] env[61728]: DEBUG nova.objects.instance [None req-0816722b-7fca-4fc5-9d34-1c1717bf3bae tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lazy-loading 'flavor' on Instance uuid 37fe7d40-cce0-41c2-be6f-d608f3503077 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1289.501105] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465162, 'name': RemoveSnapshot_Task, 'duration_secs': 0.64952} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.501477] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Deleted Snapshot of the VM instance {{(pid=61728) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1289.501806] env[61728]: DEBUG nova.compute.manager [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1289.502582] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7324b8d9-f723-45f9-8b11-dc3926a18479 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.529657] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0816722b-7fca-4fc5-9d34-1c1717bf3bae tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1289.529830] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0816722b-7fca-4fc5-9d34-1c1717bf3bae tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1289.530034] env[61728]: DEBUG nova.network.neutron [None req-0816722b-7fca-4fc5-9d34-1c1717bf3bae tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1289.530231] env[61728]: DEBUG nova.objects.instance [None req-0816722b-7fca-4fc5-9d34-1c1717bf3bae tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lazy-loading 'info_cache' on Instance uuid 37fe7d40-cce0-41c2-be6f-d608f3503077 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1290.013942] env[61728]: INFO nova.compute.manager [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Shelve offloading [ 1290.015607] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1290.015864] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e0ee6ec-c5a4-4a3c-a430-14127dc8a4a2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.023599] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1290.023599] env[61728]: value = "task-465163" [ 1290.023599] env[61728]: _type = "Task" [ 1290.023599] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.031602] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465163, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.032752] env[61728]: DEBUG nova.objects.base [None req-0816722b-7fca-4fc5-9d34-1c1717bf3bae tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Object Instance<37fe7d40-cce0-41c2-be6f-d608f3503077> lazy-loaded attributes: flavor,info_cache {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1290.534408] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] VM already powered off {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1290.534593] env[61728]: DEBUG nova.compute.manager [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1290.537111] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c396b2f-1487-402c-a7ea-c94cd3d37795 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.542784] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1290.543009] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1290.543205] env[61728]: DEBUG nova.network.neutron [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1290.734565] env[61728]: DEBUG nova.network.neutron [None req-0816722b-7fca-4fc5-9d34-1c1717bf3bae tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Updating instance_info_cache with network_info: [{"id": "ff94ce10-67c4-4c4c-baf9-d236d1670542", "address": "fa:16:3e:3d:06:bf", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff94ce10-67", "ovs_interfaceid": "ff94ce10-67c4-4c4c-baf9-d236d1670542", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1291.230285] env[61728]: DEBUG nova.network.neutron [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updating instance_info_cache with network_info: [{"id": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "address": "fa:16:3e:30:bd:13", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.237", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3aff6d0-54", "ovs_interfaceid": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1291.237608] env[61728]: DEBUG oslo_concurrency.lockutils [None req-0816722b-7fca-4fc5-9d34-1c1717bf3bae tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1291.733522] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1291.740630] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0816722b-7fca-4fc5-9d34-1c1717bf3bae tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1291.740866] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-691540e5-cc64-41ee-ae0f-81a82f39f6d5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.748751] env[61728]: DEBUG oslo_vmware.api [None req-0816722b-7fca-4fc5-9d34-1c1717bf3bae tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1291.748751] env[61728]: value = "task-465164" [ 1291.748751] env[61728]: _type = "Task" [ 1291.748751] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.756121] env[61728]: DEBUG oslo_vmware.api [None req-0816722b-7fca-4fc5-9d34-1c1717bf3bae tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465164, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.054367] env[61728]: DEBUG nova.compute.manager [req-0e946ab3-8598-43fd-8bd4-2011126c2a69 req-d9afa466-ac85-43aa-a1b1-a3da2ef5b466 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Received event network-vif-unplugged-a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1292.054602] env[61728]: DEBUG oslo_concurrency.lockutils [req-0e946ab3-8598-43fd-8bd4-2011126c2a69 req-d9afa466-ac85-43aa-a1b1-a3da2ef5b466 service nova] Acquiring lock "f95e28ce-0792-488a-a730-0766272ab2d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1292.054895] env[61728]: DEBUG oslo_concurrency.lockutils [req-0e946ab3-8598-43fd-8bd4-2011126c2a69 req-d9afa466-ac85-43aa-a1b1-a3da2ef5b466 service nova] Lock "f95e28ce-0792-488a-a730-0766272ab2d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1292.054952] env[61728]: DEBUG oslo_concurrency.lockutils [req-0e946ab3-8598-43fd-8bd4-2011126c2a69 req-d9afa466-ac85-43aa-a1b1-a3da2ef5b466 service nova] Lock "f95e28ce-0792-488a-a730-0766272ab2d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1292.055162] env[61728]: DEBUG nova.compute.manager [req-0e946ab3-8598-43fd-8bd4-2011126c2a69 req-d9afa466-ac85-43aa-a1b1-a3da2ef5b466 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] No waiting events found dispatching network-vif-unplugged-a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1292.055346] env[61728]: WARNING nova.compute.manager [req-0e946ab3-8598-43fd-8bd4-2011126c2a69 req-d9afa466-ac85-43aa-a1b1-a3da2ef5b466 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Received unexpected event network-vif-unplugged-a3aff6d0-5474-4d96-9da8-d7b907af2138 for instance with vm_state shelved and task_state shelving_offloading. [ 1292.259079] env[61728]: DEBUG oslo_vmware.api [None req-0816722b-7fca-4fc5-9d34-1c1717bf3bae tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465164, 'name': PowerOnVM_Task, 'duration_secs': 0.39557} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.259435] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-0816722b-7fca-4fc5-9d34-1c1717bf3bae tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1292.259557] env[61728]: DEBUG nova.compute.manager [None req-0816722b-7fca-4fc5-9d34-1c1717bf3bae tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1292.260323] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f80ebf0-24f3-41b1-ad8c-1cd34b0acdef {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.348126] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1292.349028] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2854541f-b93d-4bd2-8c38-0a0832e6f97b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.356254] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1292.356473] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25311ab0-d1d1-4531-a5db-9f43cba0ee20 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.418475] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1292.418715] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1292.419258] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleting the datastore file [datastore1] f95e28ce-0792-488a-a730-0766272ab2d5 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1292.419546] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a6d7a21-558e-47d1-b95b-218240486514 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.425444] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1292.425444] env[61728]: value = "task-465166" [ 1292.425444] env[61728]: _type = "Task" [ 1292.425444] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.432932] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465166, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.935177] env[61728]: DEBUG oslo_vmware.api [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465166, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161299} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.935446] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1292.935626] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1292.935808] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1292.956048] env[61728]: INFO nova.scheduler.client.report [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleted allocations for instance f95e28ce-0792-488a-a730-0766272ab2d5 [ 1293.460580] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1293.460580] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1293.460962] env[61728]: DEBUG nova.objects.instance [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lazy-loading 'resources' on Instance uuid f95e28ce-0792-488a-a730-0766272ab2d5 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1293.963926] env[61728]: DEBUG nova.objects.instance [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lazy-loading 'numa_topology' on Instance uuid f95e28ce-0792-488a-a730-0766272ab2d5 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1294.079640] env[61728]: DEBUG nova.compute.manager [req-709030a5-22ea-4b0f-9e3b-0e0637752fe5 req-34bc939c-8fa9-47b0-b58e-916d3354059f service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Received event network-changed-a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1294.079869] env[61728]: DEBUG nova.compute.manager [req-709030a5-22ea-4b0f-9e3b-0e0637752fe5 req-34bc939c-8fa9-47b0-b58e-916d3354059f service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Refreshing instance network info cache due to event network-changed-a3aff6d0-5474-4d96-9da8-d7b907af2138. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1294.080249] env[61728]: DEBUG oslo_concurrency.lockutils [req-709030a5-22ea-4b0f-9e3b-0e0637752fe5 req-34bc939c-8fa9-47b0-b58e-916d3354059f service nova] Acquiring lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1294.080431] env[61728]: DEBUG oslo_concurrency.lockutils [req-709030a5-22ea-4b0f-9e3b-0e0637752fe5 req-34bc939c-8fa9-47b0-b58e-916d3354059f service nova] Acquired lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1294.080613] env[61728]: DEBUG nova.network.neutron [req-709030a5-22ea-4b0f-9e3b-0e0637752fe5 req-34bc939c-8fa9-47b0-b58e-916d3354059f service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Refreshing network info cache for port a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1294.091798] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e153db49-484d-4cd5-b581-23445489364c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.098993] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-beca4a41-d7c2-40fe-9ce0-f162fdd5b826 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Suspending the VM {{(pid=61728) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1294.099427] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-23b51bbe-f68a-4dc2-b0bd-159ad00d0e8d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.107357] env[61728]: DEBUG oslo_vmware.api [None req-beca4a41-d7c2-40fe-9ce0-f162fdd5b826 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1294.107357] env[61728]: value = "task-465167" [ 1294.107357] env[61728]: _type = "Task" [ 1294.107357] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.115574] env[61728]: DEBUG oslo_vmware.api [None req-beca4a41-d7c2-40fe-9ce0-f162fdd5b826 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465167, 'name': SuspendVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.467352] env[61728]: DEBUG nova.objects.base [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61728) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1294.498590] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b93486-da8c-423a-a1d5-6a2d8a509c0a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.506708] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653803d5-5fdb-4081-bd17-ea63ae4e49e9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.538392] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-071535e8-24ac-4375-9807-aa3cae1b0168 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.549219] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb9381b-3d42-406b-8cd4-66163deca2db {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.559953] env[61728]: DEBUG nova.compute.provider_tree [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1294.621132] env[61728]: DEBUG oslo_vmware.api [None req-beca4a41-d7c2-40fe-9ce0-f162fdd5b826 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465167, 'name': SuspendVM_Task} progress is 58%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.849070] env[61728]: DEBUG nova.network.neutron [req-709030a5-22ea-4b0f-9e3b-0e0637752fe5 req-34bc939c-8fa9-47b0-b58e-916d3354059f service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updated VIF entry in instance network info cache for port a3aff6d0-5474-4d96-9da8-d7b907af2138. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1294.849463] env[61728]: DEBUG nova.network.neutron [req-709030a5-22ea-4b0f-9e3b-0e0637752fe5 req-34bc939c-8fa9-47b0-b58e-916d3354059f service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updating instance_info_cache with network_info: [{"id": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "address": "fa:16:3e:30:bd:13", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.237", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapa3aff6d0-54", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1295.063199] env[61728]: DEBUG nova.scheduler.client.report [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1295.118514] env[61728]: DEBUG oslo_vmware.api [None req-beca4a41-d7c2-40fe-9ce0-f162fdd5b826 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465167, 'name': SuspendVM_Task, 'duration_secs': 0.669611} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1295.118778] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-beca4a41-d7c2-40fe-9ce0-f162fdd5b826 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Suspended the VM {{(pid=61728) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1295.118968] env[61728]: DEBUG nova.compute.manager [None req-beca4a41-d7c2-40fe-9ce0-f162fdd5b826 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1295.119709] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4959d52e-9777-4993-8406-f22a59a37324 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.352057] env[61728]: DEBUG oslo_concurrency.lockutils [req-709030a5-22ea-4b0f-9e3b-0e0637752fe5 req-34bc939c-8fa9-47b0-b58e-916d3354059f service nova] Releasing lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1295.569064] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.108s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1295.571678] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "f95e28ce-0792-488a-a730-0766272ab2d5" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1296.075883] env[61728]: DEBUG oslo_concurrency.lockutils [None req-c5917949-da99-4db3-99f9-077e7085e4e9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "f95e28ce-0792-488a-a730-0766272ab2d5" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.389s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1296.076715] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "f95e28ce-0792-488a-a730-0766272ab2d5" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.505s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1296.076899] env[61728]: INFO nova.compute.manager [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Unshelving [ 1296.469321] env[61728]: INFO nova.compute.manager [None req-8a3eb93c-e9b2-4339-8793-27b37bb89128 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Resuming [ 1296.470035] env[61728]: DEBUG nova.objects.instance [None req-8a3eb93c-e9b2-4339-8793-27b37bb89128 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lazy-loading 'flavor' on Instance uuid 37fe7d40-cce0-41c2-be6f-d608f3503077 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1297.101522] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1297.101834] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1297.102091] env[61728]: DEBUG nova.objects.instance [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lazy-loading 'pci_requests' on Instance uuid f95e28ce-0792-488a-a730-0766272ab2d5 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1297.477717] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8a3eb93c-e9b2-4339-8793-27b37bb89128 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1297.477918] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8a3eb93c-e9b2-4339-8793-27b37bb89128 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquired lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1297.478160] env[61728]: DEBUG nova.network.neutron [None req-8a3eb93c-e9b2-4339-8793-27b37bb89128 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1297.605762] env[61728]: DEBUG nova.objects.instance [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lazy-loading 'numa_topology' on Instance uuid f95e28ce-0792-488a-a730-0766272ab2d5 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1298.108200] env[61728]: INFO nova.compute.claims [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1298.176088] env[61728]: DEBUG nova.network.neutron [None req-8a3eb93c-e9b2-4339-8793-27b37bb89128 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Updating instance_info_cache with network_info: [{"id": "ff94ce10-67c4-4c4c-baf9-d236d1670542", "address": "fa:16:3e:3d:06:bf", "network": {"id": "c44313df-8bb9-4dd3-849e-ac615a41c144", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-961510397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7f61c87df8b485b9a6370032dd7af37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff94ce10-67", "ovs_interfaceid": "ff94ce10-67c4-4c4c-baf9-d236d1670542", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1298.679014] env[61728]: DEBUG oslo_concurrency.lockutils [None req-8a3eb93c-e9b2-4339-8793-27b37bb89128 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Releasing lock "refresh_cache-37fe7d40-cce0-41c2-be6f-d608f3503077" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1298.679999] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b7127a4-b185-4e27-b0e9-5c9861cb2e4f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.686739] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8a3eb93c-e9b2-4339-8793-27b37bb89128 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Resuming the VM {{(pid=61728) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1298.686981] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a856823-a199-4f33-a7fe-6c7645580692 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.693281] env[61728]: DEBUG oslo_vmware.api [None req-8a3eb93c-e9b2-4339-8793-27b37bb89128 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1298.693281] env[61728]: value = "task-465168" [ 1298.693281] env[61728]: _type = "Task" [ 1298.693281] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1298.700481] env[61728]: DEBUG oslo_vmware.api [None req-8a3eb93c-e9b2-4339-8793-27b37bb89128 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465168, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1299.162073] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6410373-0c90-492e-a99f-df608db145b0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.170806] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4da306d-081a-427b-b806-ab4867c62d31 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.206166] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac8dba3-b28c-4621-9dd1-dd41f0d5b8c1 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.214865] env[61728]: DEBUG oslo_vmware.api [None req-8a3eb93c-e9b2-4339-8793-27b37bb89128 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465168, 'name': PowerOnVM_Task, 'duration_secs': 0.515993} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1299.217463] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-8a3eb93c-e9b2-4339-8793-27b37bb89128 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Resumed the VM {{(pid=61728) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1299.217830] env[61728]: DEBUG nova.compute.manager [None req-8a3eb93c-e9b2-4339-8793-27b37bb89128 tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1299.219164] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e1bcd3-6c1e-4518-995b-875dc388c34e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.224828] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f606b12-9f3a-419a-a22d-315dc0104391 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.242814] env[61728]: DEBUG nova.compute.provider_tree [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1299.750465] env[61728]: DEBUG nova.scheduler.client.report [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1300.083791] env[61728]: DEBUG oslo_concurrency.lockutils [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "37fe7d40-cce0-41c2-be6f-d608f3503077" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1300.084079] env[61728]: DEBUG oslo_concurrency.lockutils [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "37fe7d40-cce0-41c2-be6f-d608f3503077" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1300.084360] env[61728]: DEBUG oslo_concurrency.lockutils [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "37fe7d40-cce0-41c2-be6f-d608f3503077-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1300.084562] env[61728]: DEBUG oslo_concurrency.lockutils [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "37fe7d40-cce0-41c2-be6f-d608f3503077-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1300.084742] env[61728]: DEBUG oslo_concurrency.lockutils [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "37fe7d40-cce0-41c2-be6f-d608f3503077-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1300.087014] env[61728]: INFO nova.compute.manager [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Terminating instance [ 1300.089183] env[61728]: DEBUG nova.compute.manager [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1300.089390] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1300.090281] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6d8523-c76d-4578-bc98-9b52db50af33 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.098276] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1300.098510] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4d3d8f1-3cf6-4036-9d6a-5b3f26212cc7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.104568] env[61728]: DEBUG oslo_vmware.api [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1300.104568] env[61728]: value = "task-465169" [ 1300.104568] env[61728]: _type = "Task" [ 1300.104568] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.112196] env[61728]: DEBUG oslo_vmware.api [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465169, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.255454] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.153s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1300.283211] env[61728]: INFO nova.network.neutron [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updating port a3aff6d0-5474-4d96-9da8-d7b907af2138 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1300.615663] env[61728]: DEBUG oslo_vmware.api [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465169, 'name': PowerOffVM_Task, 'duration_secs': 0.19301} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1300.615663] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1300.615912] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1300.616040] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5c3eccb-66cf-48f3-b50d-10ac595e245a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.674291] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1300.674506] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1300.674692] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleting the datastore file [datastore1] 37fe7d40-cce0-41c2-be6f-d608f3503077 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1300.674964] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a4ceaa8f-d52b-4f86-ba84-0dc834605582 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.685909] env[61728]: DEBUG oslo_vmware.api [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for the task: (returnval){ [ 1300.685909] env[61728]: value = "task-465171" [ 1300.685909] env[61728]: _type = "Task" [ 1300.685909] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.693438] env[61728]: DEBUG oslo_vmware.api [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465171, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.195870] env[61728]: DEBUG oslo_vmware.api [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Task: {'id': task-465171, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165555} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1301.196198] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1301.196414] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1301.196598] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1301.196778] env[61728]: INFO nova.compute.manager [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1301.197028] env[61728]: DEBUG oslo.service.loopingcall [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1301.197235] env[61728]: DEBUG nova.compute.manager [-] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1301.197328] env[61728]: DEBUG nova.network.neutron [-] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1301.618885] env[61728]: DEBUG nova.compute.manager [req-3df75a88-8bac-4849-bb95-97ff9a2e7cc5 req-35111c92-ec61-471a-9017-ff6254a80b3e service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Received event network-vif-deleted-ff94ce10-67c4-4c4c-baf9-d236d1670542 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1301.619154] env[61728]: INFO nova.compute.manager [req-3df75a88-8bac-4849-bb95-97ff9a2e7cc5 req-35111c92-ec61-471a-9017-ff6254a80b3e service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Neutron deleted interface ff94ce10-67c4-4c4c-baf9-d236d1670542; detaching it from the instance and deleting it from the info cache [ 1301.619290] env[61728]: DEBUG nova.network.neutron [req-3df75a88-8bac-4849-bb95-97ff9a2e7cc5 req-35111c92-ec61-471a-9017-ff6254a80b3e service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1301.653359] env[61728]: DEBUG nova.compute.manager [req-08dfb577-5a5d-4a35-8a0a-450be3f4aced req-3c0a70dc-9c91-4c24-bdb9-475f021240e7 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Received event network-vif-plugged-a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1301.653359] env[61728]: DEBUG oslo_concurrency.lockutils [req-08dfb577-5a5d-4a35-8a0a-450be3f4aced req-3c0a70dc-9c91-4c24-bdb9-475f021240e7 service nova] Acquiring lock "f95e28ce-0792-488a-a730-0766272ab2d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1301.653359] env[61728]: DEBUG oslo_concurrency.lockutils [req-08dfb577-5a5d-4a35-8a0a-450be3f4aced req-3c0a70dc-9c91-4c24-bdb9-475f021240e7 service nova] Lock "f95e28ce-0792-488a-a730-0766272ab2d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1301.653359] env[61728]: DEBUG oslo_concurrency.lockutils [req-08dfb577-5a5d-4a35-8a0a-450be3f4aced req-3c0a70dc-9c91-4c24-bdb9-475f021240e7 service nova] Lock "f95e28ce-0792-488a-a730-0766272ab2d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1301.653359] env[61728]: DEBUG nova.compute.manager [req-08dfb577-5a5d-4a35-8a0a-450be3f4aced req-3c0a70dc-9c91-4c24-bdb9-475f021240e7 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] No waiting events found dispatching network-vif-plugged-a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1301.653359] env[61728]: WARNING nova.compute.manager [req-08dfb577-5a5d-4a35-8a0a-450be3f4aced req-3c0a70dc-9c91-4c24-bdb9-475f021240e7 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Received unexpected event network-vif-plugged-a3aff6d0-5474-4d96-9da8-d7b907af2138 for instance with vm_state shelved_offloaded and task_state spawning. [ 1301.734513] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1301.734660] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1301.734831] env[61728]: DEBUG nova.network.neutron [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Building network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1302.092405] env[61728]: DEBUG nova.network.neutron [-] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1302.121971] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-76b26219-9e27-43cc-925f-69839666817a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.132465] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657ecd15-dcbb-4091-9295-255acd89b0df {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.156590] env[61728]: DEBUG nova.compute.manager [req-3df75a88-8bac-4849-bb95-97ff9a2e7cc5 req-35111c92-ec61-471a-9017-ff6254a80b3e service nova] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Detach interface failed, port_id=ff94ce10-67c4-4c4c-baf9-d236d1670542, reason: Instance 37fe7d40-cce0-41c2-be6f-d608f3503077 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1302.492306] env[61728]: DEBUG nova.network.neutron [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updating instance_info_cache with network_info: [{"id": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "address": "fa:16:3e:30:bd:13", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.237", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3aff6d0-54", "ovs_interfaceid": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1302.595418] env[61728]: INFO nova.compute.manager [-] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Took 1.40 seconds to deallocate network for instance. [ 1302.994852] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1303.021587] env[61728]: DEBUG nova.virt.hardware [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-29T12:20:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='df05205d904e2bdd69c6359260fffa93',container_format='bare',created_at=2024-08-29T12:35:34Z,direct_url=,disk_format='vmdk',id=8a9a50dc-caf7-4cd3-afdb-3936725c0d38,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-175832191-shelved',owner='438de94d086e47b99cedc6e4a33471bf',properties=ImageMetaProps,protected=,size=31665152,status='active',tags=,updated_at=2024-08-29T12:35:46Z,virtual_size=,visibility=), allow threads: False {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1303.021874] env[61728]: DEBUG nova.virt.hardware [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Flavor limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1303.022057] env[61728]: DEBUG nova.virt.hardware [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Image limits 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1303.022260] env[61728]: DEBUG nova.virt.hardware [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Flavor pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1303.022417] env[61728]: DEBUG nova.virt.hardware [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Image pref 0:0:0 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1303.022571] env[61728]: DEBUG nova.virt.hardware [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61728) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1303.022782] env[61728]: DEBUG nova.virt.hardware [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1303.022964] env[61728]: DEBUG nova.virt.hardware [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1303.023161] env[61728]: DEBUG nova.virt.hardware [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Got 1 possible topologies {{(pid=61728) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1303.023333] env[61728]: DEBUG nova.virt.hardware [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1303.023508] env[61728]: DEBUG nova.virt.hardware [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61728) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1303.024415] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4e8528-a4ac-4e5b-98ca-d35a6526fdad {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.032489] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db42573-0d73-4bf4-8d6e-fc530dfd3cc7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.045754] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:bd:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f856fca-9fb5-41ea-a057-ac4193bd323d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a3aff6d0-5474-4d96-9da8-d7b907af2138', 'vif_model': 'vmxnet3'}] {{(pid=61728) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1303.053102] env[61728]: DEBUG oslo.service.loopingcall [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1303.053339] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Creating VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1303.053543] env[61728]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a5c118ae-398b-4f9a-ae11-17329d421b0a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.072018] env[61728]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1303.072018] env[61728]: value = "task-465172" [ 1303.072018] env[61728]: _type = "Task" [ 1303.072018] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.079191] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465172, 'name': CreateVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.101313] env[61728]: DEBUG oslo_concurrency.lockutils [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1303.101554] env[61728]: DEBUG oslo_concurrency.lockutils [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1303.101778] env[61728]: DEBUG nova.objects.instance [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lazy-loading 'resources' on Instance uuid 37fe7d40-cce0-41c2-be6f-d608f3503077 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1303.581393] env[61728]: DEBUG oslo_vmware.api [-] Task: {'id': task-465172, 'name': CreateVM_Task, 'duration_secs': 0.327552} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.581551] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Created VM on the ESX host {{(pid=61728) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1303.582268] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a9a50dc-caf7-4cd3-afdb-3936725c0d38" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1303.582438] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a9a50dc-caf7-4cd3-afdb-3936725c0d38" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1303.582842] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a9a50dc-caf7-4cd3-afdb-3936725c0d38" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1303.583105] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7376ddf7-7323-4cf1-b66d-a8e2e025fc51 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.587394] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1303.587394] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ded283-7b25-b81b-6eda-d432d2a20e50" [ 1303.587394] env[61728]: _type = "Task" [ 1303.587394] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.596932] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52ded283-7b25-b81b-6eda-d432d2a20e50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.647497] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13471ea-4753-4521-8077-d2c25161a51f {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.654806] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df74e02d-6a8c-491a-8241-b8923ef8b8f5 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.684593] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97938299-e2ac-4a79-a8c4-1e9daca72284 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.687921] env[61728]: DEBUG nova.compute.manager [req-33be65bc-242d-47e1-94f6-11e7493c8118 req-ceaf34f1-511c-44e7-8703-5dd9c8053004 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Received event network-changed-a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1303.688138] env[61728]: DEBUG nova.compute.manager [req-33be65bc-242d-47e1-94f6-11e7493c8118 req-ceaf34f1-511c-44e7-8703-5dd9c8053004 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Refreshing instance network info cache due to event network-changed-a3aff6d0-5474-4d96-9da8-d7b907af2138. {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1303.688377] env[61728]: DEBUG oslo_concurrency.lockutils [req-33be65bc-242d-47e1-94f6-11e7493c8118 req-ceaf34f1-511c-44e7-8703-5dd9c8053004 service nova] Acquiring lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1303.688531] env[61728]: DEBUG oslo_concurrency.lockutils [req-33be65bc-242d-47e1-94f6-11e7493c8118 req-ceaf34f1-511c-44e7-8703-5dd9c8053004 service nova] Acquired lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1303.688712] env[61728]: DEBUG nova.network.neutron [req-33be65bc-242d-47e1-94f6-11e7493c8118 req-ceaf34f1-511c-44e7-8703-5dd9c8053004 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Refreshing network info cache for port a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1303.695508] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff64b5b-87f3-4f88-aec7-d6d6913521d2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.709194] env[61728]: DEBUG nova.compute.provider_tree [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1304.097492] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a9a50dc-caf7-4cd3-afdb-3936725c0d38" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1304.097847] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Processing image 8a9a50dc-caf7-4cd3-afdb-3936725c0d38 {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1304.097969] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a9a50dc-caf7-4cd3-afdb-3936725c0d38/8a9a50dc-caf7-4cd3-afdb-3936725c0d38.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1304.098144] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a9a50dc-caf7-4cd3-afdb-3936725c0d38/8a9a50dc-caf7-4cd3-afdb-3936725c0d38.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1304.098332] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1304.098576] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4f2a862-fb43-47cd-bd4c-9b738d37ea2d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.106305] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1304.106474] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61728) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1304.107155] env[61728]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-223e1a0b-c2ff-429c-ac35-9ac9cfb9f412 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.111459] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1304.111459] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52af1eb9-53a0-0e58-1d89-ca9a9c0da94e" [ 1304.111459] env[61728]: _type = "Task" [ 1304.111459] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.119615] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]52af1eb9-53a0-0e58-1d89-ca9a9c0da94e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.211770] env[61728]: DEBUG nova.scheduler.client.report [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1304.303170] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1304.303362] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1304.383741] env[61728]: DEBUG nova.network.neutron [req-33be65bc-242d-47e1-94f6-11e7493c8118 req-ceaf34f1-511c-44e7-8703-5dd9c8053004 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updated VIF entry in instance network info cache for port a3aff6d0-5474-4d96-9da8-d7b907af2138. {{(pid=61728) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1304.384163] env[61728]: DEBUG nova.network.neutron [req-33be65bc-242d-47e1-94f6-11e7493c8118 req-ceaf34f1-511c-44e7-8703-5dd9c8053004 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updating instance_info_cache with network_info: [{"id": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "address": "fa:16:3e:30:bd:13", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.237", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3aff6d0-54", "ovs_interfaceid": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1304.621190] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Preparing fetch location {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1304.621447] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Fetch image to [datastore1] OSTACK_IMG_4a75e75a-7763-4036-9816-fcfb4c141604/OSTACK_IMG_4a75e75a-7763-4036-9816-fcfb4c141604.vmdk {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1304.621636] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Downloading stream optimized image 8a9a50dc-caf7-4cd3-afdb-3936725c0d38 to [datastore1] OSTACK_IMG_4a75e75a-7763-4036-9816-fcfb4c141604/OSTACK_IMG_4a75e75a-7763-4036-9816-fcfb4c141604.vmdk on the data store datastore1 as vApp {{(pid=61728) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1304.621813] env[61728]: DEBUG nova.virt.vmwareapi.images [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Downloading image file data 8a9a50dc-caf7-4cd3-afdb-3936725c0d38 to the ESX as VM named 'OSTACK_IMG_4a75e75a-7763-4036-9816-fcfb4c141604' {{(pid=61728) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1304.687783] env[61728]: DEBUG oslo_vmware.rw_handles [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1304.687783] env[61728]: value = "resgroup-9" [ 1304.687783] env[61728]: _type = "ResourcePool" [ 1304.687783] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1304.688095] env[61728]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-0bb80362-7f25-44be-bfb2-1ece733f9aed {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.709060] env[61728]: DEBUG oslo_vmware.rw_handles [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lease: (returnval){ [ 1304.709060] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529105c8-7d93-48bc-aa89-a5fcd5d86fee" [ 1304.709060] env[61728]: _type = "HttpNfcLease" [ 1304.709060] env[61728]: } obtained for vApp import into resource pool (val){ [ 1304.709060] env[61728]: value = "resgroup-9" [ 1304.709060] env[61728]: _type = "ResourcePool" [ 1304.709060] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1304.709378] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the lease: (returnval){ [ 1304.709378] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529105c8-7d93-48bc-aa89-a5fcd5d86fee" [ 1304.709378] env[61728]: _type = "HttpNfcLease" [ 1304.709378] env[61728]: } to be ready. {{(pid=61728) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1304.715788] env[61728]: DEBUG oslo_concurrency.lockutils [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.614s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1304.717562] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1304.717562] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529105c8-7d93-48bc-aa89-a5fcd5d86fee" [ 1304.717562] env[61728]: _type = "HttpNfcLease" [ 1304.717562] env[61728]: } is initializing. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1304.733795] env[61728]: INFO nova.scheduler.client.report [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Deleted allocations for instance 37fe7d40-cce0-41c2-be6f-d608f3503077 [ 1304.808065] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1304.808226] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Starting heal instance info cache {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1304.886838] env[61728]: DEBUG oslo_concurrency.lockutils [req-33be65bc-242d-47e1-94f6-11e7493c8118 req-ceaf34f1-511c-44e7-8703-5dd9c8053004 service nova] Releasing lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1305.217621] env[61728]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1305.217621] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529105c8-7d93-48bc-aa89-a5fcd5d86fee" [ 1305.217621] env[61728]: _type = "HttpNfcLease" [ 1305.217621] env[61728]: } is ready. {{(pid=61728) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1305.217969] env[61728]: DEBUG oslo_vmware.rw_handles [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1305.217969] env[61728]: value = "session[52c04abd-d61d-5fa5-052d-e244b3ac2a2d]529105c8-7d93-48bc-aa89-a5fcd5d86fee" [ 1305.217969] env[61728]: _type = "HttpNfcLease" [ 1305.217969] env[61728]: }. {{(pid=61728) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1305.218649] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a84ac0-bf8d-4797-a6cf-72222fcd4b81 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.225758] env[61728]: DEBUG oslo_vmware.rw_handles [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529d36d1-2f31-71e0-9f65-f57dfc0f9f3a/disk-0.vmdk from lease info. {{(pid=61728) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1305.225946] env[61728]: DEBUG oslo_vmware.rw_handles [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Creating HTTP connection to write to file with size = 31665152 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529d36d1-2f31-71e0-9f65-f57dfc0f9f3a/disk-0.vmdk. {{(pid=61728) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1305.283552] env[61728]: DEBUG oslo_concurrency.lockutils [None req-23eb520c-f664-415b-ac9c-a3a87a3b3afc tempest-ServerActionsTestJSON-1819655063 tempest-ServerActionsTestJSON-1819655063-project-member] Lock "37fe7d40-cce0-41c2-be6f-d608f3503077" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.199s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1305.288775] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6849f0b5-fb54-475c-ac43-b3c924da7283 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.310912] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1305.311112] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquired lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1305.311251] env[61728]: DEBUG nova.network.neutron [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Forcefully refreshing network info cache for instance {{(pid=61728) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1306.422072] env[61728]: DEBUG oslo_vmware.rw_handles [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Completed reading data from the image iterator. {{(pid=61728) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1306.422278] env[61728]: DEBUG oslo_vmware.rw_handles [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529d36d1-2f31-71e0-9f65-f57dfc0f9f3a/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1306.423291] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0408b87-0e4d-43f5-834c-3cc061bd91f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.430829] env[61728]: DEBUG oslo_vmware.rw_handles [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529d36d1-2f31-71e0-9f65-f57dfc0f9f3a/disk-0.vmdk is in state: ready. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1306.431041] env[61728]: DEBUG oslo_vmware.rw_handles [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529d36d1-2f31-71e0-9f65-f57dfc0f9f3a/disk-0.vmdk. {{(pid=61728) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1306.431286] env[61728]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-963b081e-8458-44f9-93c5-e142dfd41b89 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.537654] env[61728]: DEBUG nova.network.neutron [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updating instance_info_cache with network_info: [{"id": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "address": "fa:16:3e:30:bd:13", "network": {"id": "7b3e29fc-3939-4f3a-803b-84f520729a07", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-528169648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.237", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "438de94d086e47b99cedc6e4a33471bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3aff6d0-54", "ovs_interfaceid": "a3aff6d0-5474-4d96-9da8-d7b907af2138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1306.633165] env[61728]: DEBUG oslo_vmware.rw_handles [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529d36d1-2f31-71e0-9f65-f57dfc0f9f3a/disk-0.vmdk. {{(pid=61728) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1306.633756] env[61728]: INFO nova.virt.vmwareapi.images [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Downloaded image file data 8a9a50dc-caf7-4cd3-afdb-3936725c0d38 [ 1306.634345] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-886a5289-8248-4b3f-9ff2-62736d860926 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.653795] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0ac9bbf2-b45c-42a5-9410-2c60b8635387 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.690997] env[61728]: INFO nova.virt.vmwareapi.images [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] The imported VM was unregistered [ 1306.693540] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Caching image {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1306.693783] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Creating directory with path [datastore1] devstack-image-cache_base/8a9a50dc-caf7-4cd3-afdb-3936725c0d38 {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1306.694063] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c75be688-0a63-47ce-b68a-22c1d27633d0 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.705578] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Created directory with path [datastore1] devstack-image-cache_base/8a9a50dc-caf7-4cd3-afdb-3936725c0d38 {{(pid=61728) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1306.705783] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_4a75e75a-7763-4036-9816-fcfb4c141604/OSTACK_IMG_4a75e75a-7763-4036-9816-fcfb4c141604.vmdk to [datastore1] devstack-image-cache_base/8a9a50dc-caf7-4cd3-afdb-3936725c0d38/8a9a50dc-caf7-4cd3-afdb-3936725c0d38.vmdk. {{(pid=61728) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1306.706067] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-df94d9a1-9897-4067-9f3a-643fdab107a8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.712659] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1306.712659] env[61728]: value = "task-465175" [ 1306.712659] env[61728]: _type = "Task" [ 1306.712659] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.720164] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465175, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.041737] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Releasing lock "refresh_cache-f95e28ce-0792-488a-a730-0766272ab2d5" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1307.042056] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updated the network info_cache for instance {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1307.042300] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1307.042492] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1307.042691] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1307.042855] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1307.043024] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1307.043254] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1307.043434] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61728) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1307.043625] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1307.224124] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465175, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.546799] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1307.547200] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1307.547368] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1307.547558] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61728) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1307.548487] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22782ad0-53df-46a0-a203-0342689613fd {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.557219] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e917f7f1-7eb1-4166-b095-5dfcf5453f05 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.571391] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-862c4491-77c7-416e-bd9a-cb37e3d769f6 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.579716] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ab91a2-78bd-4708-9c19-78ce221007d3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.609379] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181515MB free_disk=115GB free_vcpus=48 pci_devices=None {{(pid=61728) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1307.609536] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1307.609754] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1307.723314] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465175, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.224440] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465175, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.636034] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Instance f95e28ce-0792-488a-a730-0766272ab2d5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61728) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1308.636376] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1308.636376] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1308.666612] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac28616-f156-4882-8777-cb089de4034b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.674524] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5040e860-6bc2-4465-9c29-6e5c86598b49 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.704707] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d44504f6-8389-47b9-91b9-153443bc5d6a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.712652] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c8ef92c-9083-42f8-9a08-8d1ddbb07e82 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.730066] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1308.734465] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465175, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1309.227386] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465175, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.350414} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1309.227756] env[61728]: INFO nova.virt.vmwareapi.ds_util [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_4a75e75a-7763-4036-9816-fcfb4c141604/OSTACK_IMG_4a75e75a-7763-4036-9816-fcfb4c141604.vmdk to [datastore1] devstack-image-cache_base/8a9a50dc-caf7-4cd3-afdb-3936725c0d38/8a9a50dc-caf7-4cd3-afdb-3936725c0d38.vmdk. [ 1309.227952] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Cleaning up location [datastore1] OSTACK_IMG_4a75e75a-7763-4036-9816-fcfb4c141604 {{(pid=61728) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1309.228145] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_4a75e75a-7763-4036-9816-fcfb4c141604 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1309.228398] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b29a7969-909f-4588-95a0-5fbcb05eda1e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.234481] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1309.234481] env[61728]: value = "task-465176" [ 1309.234481] env[61728]: _type = "Task" [ 1309.234481] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1309.237990] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1309.246409] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465176, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1309.743059] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1309.743059] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.133s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1309.746045] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465176, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.078271} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1309.746300] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1309.746473] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a9a50dc-caf7-4cd3-afdb-3936725c0d38/8a9a50dc-caf7-4cd3-afdb-3936725c0d38.vmdk" {{(pid=61728) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1309.746718] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a9a50dc-caf7-4cd3-afdb-3936725c0d38/8a9a50dc-caf7-4cd3-afdb-3936725c0d38.vmdk to [datastore1] f95e28ce-0792-488a-a730-0766272ab2d5/f95e28ce-0792-488a-a730-0766272ab2d5.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1309.746972] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6304176a-e9c3-4c7c-b01a-b017c55f01b9 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.753941] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1309.753941] env[61728]: value = "task-465177" [ 1309.753941] env[61728]: _type = "Task" [ 1309.753941] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1309.761368] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465177, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.264368] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465177, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.765884] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465177, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.266273] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465177, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.767769] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465177, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.270264] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465177, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.494609} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.270667] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a9a50dc-caf7-4cd3-afdb-3936725c0d38/8a9a50dc-caf7-4cd3-afdb-3936725c0d38.vmdk to [datastore1] f95e28ce-0792-488a-a730-0766272ab2d5/f95e28ce-0792-488a-a730-0766272ab2d5.vmdk {{(pid=61728) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1312.271882] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a5ccd0c-81fb-492b-baaf-90541d468c16 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.300688] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Reconfiguring VM instance instance-0000007c to attach disk [datastore1] f95e28ce-0792-488a-a730-0766272ab2d5/f95e28ce-0792-488a-a730-0766272ab2d5.vmdk or device None with type streamOptimized {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1312.301025] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6372e0dc-e183-40a4-86bb-98fd4ba6dff7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.320182] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1312.320182] env[61728]: value = "task-465178" [ 1312.320182] env[61728]: _type = "Task" [ 1312.320182] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.328072] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465178, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.829892] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465178, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.330109] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465178, 'name': ReconfigVM_Task, 'duration_secs': 0.619271} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1313.330397] env[61728]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Reconfigured VM instance instance-0000007c to attach disk [datastore1] f95e28ce-0792-488a-a730-0766272ab2d5/f95e28ce-0792-488a-a730-0766272ab2d5.vmdk or device None with type streamOptimized {{(pid=61728) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1313.331044] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e81d1dcb-10f9-4478-afbf-517569b3c565 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.337634] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1313.337634] env[61728]: value = "task-465179" [ 1313.337634] env[61728]: _type = "Task" [ 1313.337634] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.344931] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465179, 'name': Rename_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.847895] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465179, 'name': Rename_Task, 'duration_secs': 0.129448} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1313.848271] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Powering on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1313.848476] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8b39da89-9447-4267-b58e-74bd58f62259 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.854488] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1313.854488] env[61728]: value = "task-465180" [ 1313.854488] env[61728]: _type = "Task" [ 1313.854488] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.861250] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465180, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.364155] env[61728]: DEBUG oslo_vmware.api [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465180, 'name': PowerOnVM_Task, 'duration_secs': 0.463413} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.364462] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Powered on the VM {{(pid=61728) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1314.467096] env[61728]: DEBUG nova.compute.manager [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Checking state {{(pid=61728) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1314.468213] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5922e67-3778-43c2-ad28-64138506df6e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.984627] env[61728]: DEBUG oslo_concurrency.lockutils [None req-bf5ca5d0-8635-4b9d-9413-0338eab0daf9 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "f95e28ce-0792-488a-a730-0766272ab2d5" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.908s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1347.855023] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1349.357203] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1350.849339] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1350.853984] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1351.854594] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1351.854883] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1352.241073] env[61728]: DEBUG oslo_concurrency.lockutils [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "f95e28ce-0792-488a-a730-0766272ab2d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1352.241316] env[61728]: DEBUG oslo_concurrency.lockutils [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "f95e28ce-0792-488a-a730-0766272ab2d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1352.241536] env[61728]: DEBUG oslo_concurrency.lockutils [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "f95e28ce-0792-488a-a730-0766272ab2d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1352.241723] env[61728]: DEBUG oslo_concurrency.lockutils [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "f95e28ce-0792-488a-a730-0766272ab2d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1352.241947] env[61728]: DEBUG oslo_concurrency.lockutils [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "f95e28ce-0792-488a-a730-0766272ab2d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1352.245746] env[61728]: INFO nova.compute.manager [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Terminating instance [ 1352.247506] env[61728]: DEBUG nova.compute.manager [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Start destroying the instance on the hypervisor. {{(pid=61728) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1352.247711] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Destroying instance {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1352.248568] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f47abd7-adea-455f-88b5-6e9864ab216b {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.256929] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Powering off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1352.257167] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3807d4c6-9d74-479e-a8b2-722a7c4c50ce {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.263406] env[61728]: DEBUG oslo_vmware.api [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1352.263406] env[61728]: value = "task-465181" [ 1352.263406] env[61728]: _type = "Task" [ 1352.263406] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1352.270734] env[61728]: DEBUG oslo_vmware.api [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465181, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1352.772864] env[61728]: DEBUG oslo_vmware.api [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465181, 'name': PowerOffVM_Task, 'duration_secs': 0.18283} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1352.773175] env[61728]: DEBUG nova.virt.vmwareapi.vm_util [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Powered off the VM {{(pid=61728) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1352.773362] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Unregistering the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1352.773611] env[61728]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9cf9eed7-4803-4ce0-b311-ab538cfe1782 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.830624] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Unregistered the VM {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1352.830869] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Deleting contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1352.831086] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleting the datastore file [datastore1] f95e28ce-0792-488a-a730-0766272ab2d5 {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1352.831365] env[61728]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f25dc8dc-c890-49b8-871b-c3aefb776f4e {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.837960] env[61728]: DEBUG oslo_vmware.api [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for the task: (returnval){ [ 1352.837960] env[61728]: value = "task-465183" [ 1352.837960] env[61728]: _type = "Task" [ 1352.837960] env[61728]: } to complete. {{(pid=61728) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1352.845685] env[61728]: DEBUG oslo_vmware.api [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465183, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1353.348275] env[61728]: DEBUG oslo_vmware.api [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Task: {'id': task-465183, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136035} completed successfully. {{(pid=61728) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1353.348735] env[61728]: DEBUG nova.virt.vmwareapi.ds_util [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleted the datastore file {{(pid=61728) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1353.348735] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Deleted contents of the VM from datastore datastore1 {{(pid=61728) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1353.348905] env[61728]: DEBUG nova.virt.vmwareapi.vmops [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Instance destroyed {{(pid=61728) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1353.349094] env[61728]: INFO nova.compute.manager [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1353.349352] env[61728]: DEBUG oslo.service.loopingcall [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61728) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1353.349556] env[61728]: DEBUG nova.compute.manager [-] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Deallocating network for instance {{(pid=61728) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1353.349648] env[61728]: DEBUG nova.network.neutron [-] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] deallocate_for_instance() {{(pid=61728) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1353.739894] env[61728]: DEBUG nova.compute.manager [req-b4149f45-5284-4d78-b8ce-7bb4ff2969ea req-4158dd82-fcbb-4f06-a6a7-375f56eda7f3 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Received event network-vif-deleted-a3aff6d0-5474-4d96-9da8-d7b907af2138 {{(pid=61728) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1353.740039] env[61728]: INFO nova.compute.manager [req-b4149f45-5284-4d78-b8ce-7bb4ff2969ea req-4158dd82-fcbb-4f06-a6a7-375f56eda7f3 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Neutron deleted interface a3aff6d0-5474-4d96-9da8-d7b907af2138; detaching it from the instance and deleting it from the info cache [ 1353.740281] env[61728]: DEBUG nova.network.neutron [req-b4149f45-5284-4d78-b8ce-7bb4ff2969ea req-4158dd82-fcbb-4f06-a6a7-375f56eda7f3 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1353.853935] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1353.854072] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Starting heal instance info cache {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1353.854244] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Rebuilding the list of instances to heal {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1354.212367] env[61728]: DEBUG nova.network.neutron [-] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Updating instance_info_cache with network_info: [] {{(pid=61728) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1354.242808] env[61728]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f930343b-e347-4708-ac5b-071a6eb49f68 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.252817] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1254a3b5-c56d-441f-9362-88b377b70343 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.275738] env[61728]: DEBUG nova.compute.manager [req-b4149f45-5284-4d78-b8ce-7bb4ff2969ea req-4158dd82-fcbb-4f06-a6a7-375f56eda7f3 service nova] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Detach interface failed, port_id=a3aff6d0-5474-4d96-9da8-d7b907af2138, reason: Instance f95e28ce-0792-488a-a730-0766272ab2d5 could not be found. {{(pid=61728) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1354.357477] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Skipping network cache update for instance because it is being deleted. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 1354.357731] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Didn't find any instances for network info cache update. {{(pid=61728) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1354.357872] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1354.358051] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1354.715341] env[61728]: INFO nova.compute.manager [-] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Took 1.37 seconds to deallocate network for instance. [ 1354.861330] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1354.861558] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1354.861729] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1354.861888] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61728) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1354.862895] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce50fef4-967a-4c90-a175-d1808534d51a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.871422] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bcdee3e-058e-4c3a-8494-63a9b53a349c {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.885201] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8b193a-316d-4b92-9a8e-d1c9c043a997 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.891224] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1349d2-3f52-44be-b8c8-3a9a062244f4 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.585324] env[61728]: DEBUG oslo_concurrency.lockutils [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1355.585794] env[61728]: DEBUG oslo_concurrency.lockutils [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1355.585794] env[61728]: DEBUG nova.objects.instance [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lazy-loading 'resources' on Instance uuid f95e28ce-0792-488a-a730-0766272ab2d5 {{(pid=61728) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1355.587664] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181515MB free_disk=115GB free_vcpus=48 pci_devices=None {{(pid=61728) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1355.587808] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1356.239717] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f16973f3-d598-473a-8362-9b29289ff9d8 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.247375] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d22f91-fea8-4d89-8267-78d6e0acc9e3 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.275436] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a701a95a-fa18-42cb-b177-c18b72eb99da {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.282012] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83656327-6746-4c76-ab80-10a4041eec8d {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.294255] env[61728]: DEBUG nova.compute.provider_tree [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1356.797140] env[61728]: DEBUG nova.scheduler.client.report [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1357.302266] env[61728]: DEBUG oslo_concurrency.lockutils [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.716s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1357.304854] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.717s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1357.323642] env[61728]: INFO nova.scheduler.client.report [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Deleted allocations for instance f95e28ce-0792-488a-a730-0766272ab2d5 [ 1357.831242] env[61728]: DEBUG oslo_concurrency.lockutils [None req-725212e9-62a8-43a1-aad5-0cfd18c894c3 tempest-AttachVolumeShelveTestJSON-1062859602 tempest-AttachVolumeShelveTestJSON-1062859602-project-member] Lock "f95e28ce-0792-488a-a730-0766272ab2d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.590s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1358.323623] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1358.323796] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61728) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1358.336425] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d35b461-e051-4293-8d43-996ba96caeb7 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.344153] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca1aba5-3e46-4bec-9a36-887bfd0f1f02 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.373019] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a78ddcc0-5840-4b44-9178-66488c443aa2 {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.379866] env[61728]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8b6653-f6a2-470d-9756-9466dd187b6a {{(pid=61728) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.392239] env[61728]: DEBUG nova.compute.provider_tree [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed in ProviderTree for provider: e7ceb92f-072b-409e-b888-6fe0676b32f1 {{(pid=61728) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1358.895360] env[61728]: DEBUG nova.scheduler.client.report [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Inventory has not changed for provider e7ceb92f-072b-409e-b888-6fe0676b32f1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 115, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61728) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1359.400018] env[61728]: DEBUG nova.compute.resource_tracker [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61728) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1359.400265] env[61728]: DEBUG oslo_concurrency.lockutils [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.095s {{(pid=61728) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1359.400514] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1359.400686] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Cleaning up deleted instances {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1359.911290] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] There are 28 instances to clean {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1359.911670] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 37fe7d40-cce0-41c2-be6f-d608f3503077] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1360.415273] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f95e28ce-0792-488a-a730-0766272ab2d5] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1360.919060] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: e5943edd-bdc3-4285-b2e0-7adac9acaf7f] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1361.422474] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 6218d586-0190-4a02-ae88-3f74329dc32c] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1361.925919] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 76429588-fb4b-4215-b78c-75ce366989ed] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1362.429252] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: fb1195b2-1ab5-47d7-be86-4b61e47b4e29] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1362.933903] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: c92e595d-c686-4810-bb39-d98de4141121] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1363.437971] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 2db96467-5fb4-49f4-954b-dbdd21e46232] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1363.941333] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: dbe877e6-1754-4cf7-b454-abd88bfa713a] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1364.446070] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 79782d2b-1b90-4f3f-a1ce-394e88a00546] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1364.949056] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 8ea37e93-16fc-4c60-9949-17656218d46a] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1365.452533] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: c1ed0e8b-5cf9-4ff0-931e-fadc3432a125] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1365.955833] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 5eb3c3fa-d612-43d8-875c-2ee595e29081] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1366.459428] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 32a781b0-2db4-4883-ae96-d94c8181902b] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1366.963213] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 532507fa-996d-45c7-bf70-f09c93be79ed] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1367.466948] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 3c6170e8-21e5-4e6a-b5c9-402b83fc7ea4] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1367.970288] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: e7f4de3d-20fc-45a5-9b87-8d62f1bd5ed0] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1368.474305] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 77b8f509-17e7-4c25-a43d-1655c16ce8d1] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1368.978069] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 41329d0d-3838-4d1b-a4e6-8df4508eea76] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1369.481978] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 6c154517-68e8-4e70-86eb-cf53f6f812d8] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1369.985083] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: d55beb02-00fc-4df6-a239-9e5d776a7c18] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1370.488962] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 267880ff-0fac-4763-82be-ef8587ee7628] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1370.993304] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 4a5fbe5b-71e0-4e2d-9626-da8c73ccac2a] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1371.496643] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 8c5b2380-bd14-451e-9612-2767c2a6adb2] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1371.999960] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: b1fb2e3f-234a-44c7-983e-d4441b4e3cef] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1372.503190] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 5fa2e9f4-6d68-46c2-a549-95d216bab886] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1373.006906] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: f27f22aa-f05d-4981-a389-311731c51f93] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1373.510320] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] [instance: 40ef29df-12d6-4274-846a-4a2c32752b8b] Instance has had 0 of 5 cleanup attempts {{(pid=61728) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1374.013723] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1374.014164] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Cleaning up deleted instances with incomplete migration {{(pid=61728) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1375.013310] env[61728]: DEBUG oslo_service.periodic_task [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61728) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1375.013500] env[61728]: DEBUG nova.compute.manager [None req-1bbff92c-7953-4d15-b679-432894d87409 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61728) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}}